.Guarantee compatibility with numerous platforms, including.NET 6.0,. Internet Platform 4.6.2, and.NET Requirement 2.0 and also above.Minimize reliances to stop model disputes and the need for tiing redirects.Transcribing Audio Data.Among the primary functions of the SDK is audio transcription. Programmers can record audio files asynchronously or even in real-time. Below is actually an example of how to translate an audio documents:.using AssemblyAI.making use of AssemblyAI.Transcripts.var client = brand new AssemblyAIClient(" YOUR_API_KEY").var records = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For nearby reports, similar code may be utilized to obtain transcription.await utilizing var flow = brand-new FileStream("./ nbc.mp3", FileMode.Open).var transcript = await client.Transcripts.TranscribeAsync(.flow,.brand new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK additionally reinforces real-time sound transcription utilizing Streaming Speech-to-Text. This function is actually specifically valuable for treatments demanding urgent processing of audio records.utilizing AssemblyAI.Realtime.await using var transcriber = new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Ultimate: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for obtaining sound from a mic as an example.GetAudio( async (portion) => await transcriber.SendAudioAsync( part)).await transcriber.CloseAsync().Utilizing LeMUR for LLM Applications.The SDK incorporates along with LeMUR to make it possible for designers to create large foreign language version (LLM) apps on vocal data. Listed below is actually an example:.var lemurTaskParams = brand-new LemurTaskParams.Prompt="Deliver a brief rundown of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var action = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Intelligence Styles.Furthermore, the SDK possesses integrated support for audio knowledge designs, allowing conviction review as well as various other innovative features.var transcript = wait for client.Transcripts.TranscribeAsync( brand-new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = accurate. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// GOOD, NEUTRAL, or even NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For more information, go to the formal AssemblyAI blog.Image resource: Shutterstock.