.Ensure compatibility with several structures, including.NET 6.0,. Web Platform 4.6.2, and.NET Standard 2.0 and above.Minimize addictions to stop variation disputes as well as the necessity for binding redirects.Transcribing Sound Record.One of the main performances of the SDK is actually audio transcription. Developers may translate audio files asynchronously or in real-time. Below is an example of how to transcribe an audio data:.utilizing AssemblyAI.using AssemblyAI.Transcripts.var client = brand-new AssemblyAIClient(" YOUR_API_KEY").var transcript = await client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For local files, identical code could be used to obtain transcription.await using var stream = new FileStream("./ nbc.mp3", FileMode.Open).var transcript = await client.Transcripts.TranscribeAsync(.flow,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK also reinforces real-time sound transcription using Streaming Speech-to-Text. This function is especially beneficial for uses demanding quick handling of audio data.making use of AssemblyAI.Realtime.wait for using var transcriber = brand new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Ultimate: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for receiving sound from a microphone as an example.GetAudio( async (piece) => await transcriber.SendAudioAsync( chunk)).wait for transcriber.CloseAsync().Making Use Of LeMUR for LLM Apps.The SDK incorporates along with LeMUR to enable developers to create large foreign language model (LLM) functions on voice data. Listed here is actually an instance:.var lemurTaskParams = new LemurTaskParams.Motivate="Deliver a short rundown of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var response = wait for client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Intelligence Designs.Also, the SDK comes with integrated support for audio cleverness designs, making it possible for belief study as well as other advanced attributes.var transcript = wait for client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = real. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// GOOD, NEUTRAL, or even downside.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To find out more, see the official AssemblyAI blog.Image source: Shutterstock.