.Make sure compatibility along with several structures, including.NET 6.0,. Web Platform 4.6.2, and.NET Criterion 2.0 as well as above.Decrease dependences to avoid version disputes and the requirement for tiing redirects.Translating Sound Files.Some of the major capabilities of the SDK is audio transcription. Developers may transcribe audio reports asynchronously or even in real-time. Below is an example of exactly how to translate an audio data:.making use of AssemblyAI.using AssemblyAI.Transcripts.var client = brand new AssemblyAIClient(" YOUR_API_KEY").var transcript = await client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For nearby reports, similar code could be used to attain transcription.wait for using var flow = new FileStream("./ nbc.mp3", FileMode.Open).var transcript = await client.Transcripts.TranscribeAsync(.stream,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK additionally sustains real-time audio transcription utilizing Streaming Speech-to-Text. This attribute is particularly practical for requests needing instant handling of audio records.making use of AssemblyAI.Realtime.wait for utilizing var transcriber = brand new RealtimeTranscriber( brand new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Ultimate: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for receiving audio from a microphone as an example.GetAudio( async (portion) => await transcriber.SendAudioAsync( portion)).wait for transcriber.CloseAsync().Making Use Of LeMUR for LLM Functions.The SDK includes with LeMUR to enable developers to create big foreign language version (LLM) functions on vocal records. Right here is an example:.var lemurTaskParams = brand-new LemurTaskParams.Urge="Supply a brief rundown of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var reaction = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Intelligence Designs.Additionally, the SDK possesses built-in support for audio knowledge styles, enabling feeling analysis as well as other enhanced functions.var transcript = wait for client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = true. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// BENEFICIAL, NEUTRAL, or even downside.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For more information, see the main AssemblyAI blog.Image source: Shutterstock.