.Make certain compatibility along with multiple structures, including.NET 6.0,. NET Platform 4.6.2, and.NET Requirement 2.0 and also above.Lessen dependencies to prevent version disagreements and also the demand for tiing redirects.Recording Sound Data.One of the primary performances of the SDK is audio transcription. Designers may translate audio reports asynchronously or even in real-time. Below is actually an instance of how to translate an audio file:.utilizing AssemblyAI.utilizing AssemblyAI.Transcripts.var customer = brand new AssemblyAIClient(" YOUR_API_KEY").var transcript = wait for client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For neighborhood files, comparable code may be used to achieve transcription.wait for making use of var flow = new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.flow,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK likewise reinforces real-time audio transcription making use of Streaming Speech-to-Text. This feature is particularly valuable for applications calling for urgent handling of audio records.utilizing AssemblyAI.Realtime.wait for making use of var scribe = brand new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Last: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for acquiring audio coming from a microphone as an example.GetAudio( async (portion) => await transcriber.SendAudioAsync( chunk)).wait for transcriber.CloseAsync().Using LeMUR for LLM Applications.The SDK includes with LeMUR to permit programmers to construct large foreign language style (LLM) functions on vocal data. Listed here is an instance:.var lemurTaskParams = new LemurTaskParams.Motivate="Give a brief conclusion of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var feedback = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Knowledge Models.Also, the SDK includes integrated support for audio knowledge styles, making it possible for sentiment evaluation as well as various other advanced functions.var transcript = await client.Transcripts.TranscribeAsync( brand-new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = correct. ).foreach (var cause transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// FAVORABLE, NEUTRAL, or even NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To learn more, explore the official AssemblyAI blog.Image resource: Shutterstock.