.Guarantee being compatible along with numerous structures, including.NET 6.0,. NET Framework 4.6.2, and.NET Requirement 2.0 as well as above.Reduce dependencies to avoid model disagreements and also the necessity for binding redirects.Translating Sound Files.Some of the main performances of the SDK is actually audio transcription. Designers can translate audio reports asynchronously or in real-time. Below is actually an example of exactly how to record an audio data:.making use of AssemblyAI.making use of AssemblyAI.Transcripts.var client = brand new AssemblyAIClient(" YOUR_API_KEY").var records = wait for client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For nearby documents, comparable code may be used to attain transcription.wait for using var stream = new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.flow,.brand new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK additionally supports real-time audio transcription making use of Streaming Speech-to-Text. This component is actually particularly helpful for uses requiring quick handling of audio data.using AssemblyAI.Realtime.wait for using var transcriber = brand new RealtimeTranscriber( brand-new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Ultimate: transcript.Text "). ).await transcriber.ConnectAsync().// Pseudocode for obtaining audio coming from a microphone as an example.GetAudio( async (portion) => wait for transcriber.SendAudioAsync( portion)).wait for transcriber.CloseAsync().Utilizing LeMUR for LLM Applications.The SDK incorporates along with LeMUR to allow creators to create sizable foreign language style (LLM) functions on vocal information. Listed here is actually an instance:.var lemurTaskParams = new LemurTaskParams.Trigger="Provide a short recap of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var feedback = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Knowledge Styles.In addition, the SDK includes integrated help for audio intellect designs, making it possible for feeling analysis as well as other advanced attributes.var records = await client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = correct. ).foreach (var lead to transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// POSITIVE, NEUTRAL, or even NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To find out more, visit the main AssemblyAI blog.Image resource: Shutterstock.