.Ensure compatibility along with a number of platforms, including.NET 6.0,. Internet Framework 4.6.2, and.NET Requirement 2.0 and above.Reduce dependencies to stop model disputes and the requirement for binding redirects.Recording Sound Data.Among the main capabilities of the SDK is actually audio transcription. Designers can transcribe audio files asynchronously or in real-time. Below is actually an example of how to transcribe an audio report:.using AssemblyAI.utilizing AssemblyAI.Transcripts.var customer = brand new AssemblyAIClient(" YOUR_API_KEY").var transcript = wait for client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For local area data, identical code can be utilized to obtain transcription.await making use of var stream = new FileStream("./ nbc.mp3", FileMode.Open).var transcript = await client.Transcripts.TranscribeAsync(.flow,.new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK also reinforces real-time sound transcription making use of Streaming Speech-to-Text. This attribute is specifically practical for uses demanding quick processing of audio information.using AssemblyAI.Realtime.wait for using var transcriber = brand new RealtimeTranscriber( brand-new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Ultimate: transcript.Text "). ).await transcriber.ConnectAsync().// Pseudocode for obtaining sound coming from a mic for instance.GetAudio( async (part) => wait for transcriber.SendAudioAsync( piece)).wait for transcriber.CloseAsync().Using LeMUR for LLM Applications.The SDK incorporates with LeMUR to allow programmers to develop huge foreign language design (LLM) functions on vocal data. Listed here is actually an instance:.var lemurTaskParams = brand-new LemurTaskParams.Trigger="Give a quick rundown of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var action = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Intelligence Styles.In addition, the SDK features integrated assistance for audio cleverness versions, permitting view study and also various other sophisticated components.var transcript = wait for client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = true. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// FAVORABLE, NEUTRAL, or downside.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For additional information, go to the formal AssemblyAI blog.Image source: Shutterstock.