.Guarantee compatibility along with multiple platforms, including.NET 6.0,. NET Framework 4.6.2, and.NET Criterion 2.0 and above.Reduce reliances to avoid version disagreements and the demand for tiing redirects.Translating Sound Information.One of the major functionalities of the SDK is audio transcription. Creators can record audio reports asynchronously or even in real-time. Below is actually an example of how to translate an audio documents:.using AssemblyAI.utilizing AssemblyAI.Transcripts.var client = brand new AssemblyAIClient(" YOUR_API_KEY").var transcript = await client.Transcripts.TranscribeAsync( brand-new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For neighborhood files, identical code may be used to achieve transcription.await utilizing var stream = new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.stream,.new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK also supports real-time audio transcription using Streaming Speech-to-Text. This attribute is specifically useful for applications requiring instant processing of audio records.making use of AssemblyAI.Realtime.await using var transcriber = brand-new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Last: transcript.Text "). ).await transcriber.ConnectAsync().// Pseudocode for receiving sound from a microphone as an example.GetAudio( async (piece) => wait for transcriber.SendAudioAsync( portion)).wait for transcriber.CloseAsync().Making Use Of LeMUR for LLM Applications.The SDK incorporates along with LeMUR to make it possible for designers to develop sizable language version (LLM) applications on voice information. Below is actually an example:.var lemurTaskParams = brand new LemurTaskParams.Cause="Deliver a brief summary of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var response = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Intellect Styles.In addition, the SDK features built-in assistance for audio intellect styles, enabling belief review and also various other innovative functions.var transcript = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = correct. ).foreach (var lead to transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// BENEFICIAL, NEUTRAL, or even NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For more details, go to the formal AssemblyAI blog.Image source: Shutterstock.