.Ensure being compatible with multiple frameworks, including.NET 6.0,. Internet Framework 4.6.2, and.NET Requirement 2.0 and also above.Lessen dependences to avoid version conflicts and also the demand for tiing redirects.Translating Audio Information.One of the primary functions of the SDK is actually audio transcription. Programmers can transcribe audio data asynchronously or in real-time. Below is actually an example of exactly how to translate an audio file:.utilizing AssemblyAI.utilizing AssemblyAI.Transcripts.var customer = brand-new AssemblyAIClient(" YOUR_API_KEY").var transcript = await client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For local area reports, comparable code could be used to achieve transcription.wait for making use of var flow = brand new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.stream,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK also holds real-time sound transcription utilizing Streaming Speech-to-Text. This feature is actually specifically valuable for requests calling for instant processing of audio information.making use of AssemblyAI.Realtime.await utilizing var scribe = brand-new RealtimeTranscriber( brand new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Last: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for obtaining sound from a mic as an example.GetAudio( async (portion) => await transcriber.SendAudioAsync( chunk)).await transcriber.CloseAsync().Utilizing LeMUR for LLM Applications.The SDK integrates along with LeMUR to permit programmers to create big language design (LLM) applications on vocal information. Here is an example:.var lemurTaskParams = brand-new LemurTaskParams.Urge="Offer a brief summary of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var feedback = wait for client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Knowledge Styles.In addition, the SDK features built-in support for audio intellect versions, making it possible for conviction evaluation and various other enhanced components.var records = wait for client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = correct. ).foreach (var lead to transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// GOOD, NEUTRAL, or NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To read more, see the main AssemblyAI blog.Image resource: Shutterstock.