.Make sure compatibility with a number of platforms, including.NET 6.0,. Internet Structure 4.6.2, and.NET Requirement 2.0 as well as above.Decrease dependences to stop version disagreements and also the need for binding redirects.Recording Sound Files.One of the major functionalities of the SDK is audio transcription. Creators may record audio files asynchronously or even in real-time. Below is an example of how to record an audio report:.making use of AssemblyAI.making use of AssemblyAI.Transcripts.var client = brand new AssemblyAIClient(" YOUR_API_KEY").var transcript = await client.Transcripts.TranscribeAsync( brand-new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For local documents, similar code can be made use of to accomplish transcription.wait for using var stream = brand-new FileStream("./ nbc.mp3", FileMode.Open).var transcript = await client.Transcripts.TranscribeAsync(.stream,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK likewise holds real-time sound transcription using Streaming Speech-to-Text. This attribute is especially useful for applications requiring urgent handling of audio records.using AssemblyAI.Realtime.wait for utilizing var transcriber = brand new RealtimeTranscriber( brand-new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Final: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for acquiring sound from a microphone as an example.GetAudio( async (portion) => wait for transcriber.SendAudioAsync( part)).await transcriber.CloseAsync().Utilizing LeMUR for LLM Applications.The SDK integrates along with LeMUR to make it possible for creators to create huge language style (LLM) applications on vocal data. Listed here is actually an example:.var lemurTaskParams = brand-new LemurTaskParams.Urge="Give a quick recap of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var action = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Cleverness Designs.In addition, the SDK features integrated help for audio knowledge styles, making it possible for conviction study as well as other advanced functions.var transcript = wait for client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = true. ).foreach (var cause transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// BENEFICIAL, NEUTRAL, or NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For more information, see the formal AssemblyAI blog.Image resource: Shutterstock.