.Ensure being compatible with a number of platforms, including.NET 6.0,. Web Structure 4.6.2, and.NET Standard 2.0 as well as above.Lessen dependencies to prevent variation disagreements and also the necessity for tiing redirects.Transcribing Audio Information.Some of the major functionalities of the SDK is audio transcription. Creators can transcribe audio reports asynchronously or even in real-time. Below is an instance of just how to translate an audio data:.using AssemblyAI.making use of AssemblyAI.Transcripts.var customer = new AssemblyAIClient(" YOUR_API_KEY").var records = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For neighborhood data, comparable code may be used to obtain transcription.await utilizing var flow = brand-new FileStream("./ nbc.mp3", FileMode.Open).var records = wait for client.Transcripts.TranscribeAsync(.stream,.brand new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK likewise reinforces real-time sound transcription using Streaming Speech-to-Text. This component is specifically useful for treatments demanding instant handling of audio records.making use of AssemblyAI.Realtime.await making use of var scribe = new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Last: transcript.Text "). ).await transcriber.ConnectAsync().// Pseudocode for acquiring sound coming from a mic for example.GetAudio( async (piece) => wait for transcriber.SendAudioAsync( piece)).await transcriber.CloseAsync().Utilizing LeMUR for LLM Apps.The SDK integrates with LeMUR to allow developers to construct large language design (LLM) apps on voice data. Listed here is an instance:.var lemurTaskParams = brand-new LemurTaskParams.Urge="Supply a short conclusion of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var action = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Intellect Models.Also, the SDK possesses integrated help for audio intellect designs, permitting belief study and also various other enhanced components.var records = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = accurate. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// BENEFICIAL, NEUTRAL, or downside.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For more details, see the formal AssemblyAI blog.Image resource: Shutterstock.