Automatic language detection improvements: increased accuracy & expanded language support
Announcing several improvements to our automatic language detection model, yielding a more accurate model that's available in more languages.
New releases, model updates, awards, events, and more.
Announcements
Announcing several improvements to our automatic language detection model, yielding a more accurate model that's available in more languages.
Announcements
The enhanced AssemblyAI app for Zapier is more flexible, supports more AssemblyAI features, and integrates more closely into the Zap editor.
Announcements
We're announcing support for 47 new languages and 16 new entities to make our Audio Intelligence more powerful and globally accessible.
Announcements
Today we're releasing Anthropic's Claude 3 model family into LeMUR as part of our ongoing commitment to giving you access to the most advanced and innovative AI capabilities available on the market.
Announcements
Announcing several improvements to our Speaker Diarization service, yielding a more accurate model that's available in more languages.
Announcements
Last month, we introduced the Best and Nano tiers — new pricing options for Speech-to-Text models to help you balance accuracy, cost, and speed. Here's a breakdown of our Best and Nano tiers.
Announcements
We've been working with Make to make our Speech AI available to no-code builders. With the AssemblyAI app for Make, you can use our speech recognition models, audio intelligence models, and build generative features with LLMs.
Announcements
Streaming Speech-to-Text makes it easier than ever to transcribe live audio and videos, now with customizable end-of-utterance detection at a lower cost.
Announcements
Learn how to build a real-world application working with audio data in this free video course.
Announcements
Learn about the improvements in the recently released v1.3.0 of the AssemblyAI Go SDK.
Announcements
We're excited to announce the release of our AssemblyAI SDK for Go!
Announcements
Transcribe, summarize, or extract speakers from your audio files using the new AssemblyAI Haystack integration.
Announcements
We’re excited to introduce major improvements to our API’s inference latency, with the majority of audio files now completing in well under 45 seconds regardless of audio duration.
Announcements
We are thrilled to release the AssemblyAI Java SDK. You can use the SDK to transcribe audio asynchronously or in real-time, use our audio intelligence model, and apply LLMs to your audio data using LeMUR.
Announcements
You can now integrate spoken audio data into Semantic Kernel .NET applications using the new AssemblyAI integration.
Announcements
Transcribe audio files in your LlamaIndex.TS applications using the new AssemblyAI speech-to-text integration.
Announcements
We're excited to share that we’ve raised $50M in Series C funding led by Accel, our partners that also led our Series A, with participation from Keith Block and Smith Point Capital, Insight Partners, Daniel Gross and Nat Friedman, and Y Combinator.
Announcements
AssemblyAI is now an official partner on the Amazon Web Services (AWS) Marketplace.
Announcements
We’ve trained new Punctuation and Truecasing models on 13 billion words to achieve a 39% F1 score improvement for mixed-case words. Building on a novel application of a hybrid architecture for a character-level classifier reduces inference time and improves the scalability of our Speech AI systems.
Announcements
We're thrilled to release version 2.0 of the AssemblyAI Node SDK. The SDK has been rewritten from scratch using TypeScript and provides easy-to-use access to our API.
Announcements
AssemblyAI transcription and LeMUR capabilities are part of Rivet, the IDE for developing AI applications.
Announcements
You can now integrate spoken audio data into LangChain.js applications using the new AssemblyAI integration.
Launch
LeMUR is the easiest way to build LLM apps on spoken data - search, summarize, and ask questions, with knowledge of your spoken data.
Launch
We're introducing Conformer-2, our latest AI model for automatic speech recognition. Conformer-2 is trained on 1.1M hours of English audio data, extending Conformer-1 to provide improvements on proper nouns, alphanumerics, and robustness to noise.
Launch
LeMUR is our new framework for applying powerful LLMs to transcribed speech. With a single line of code, LeMUR can quickly process audio transcripts for up to 10 hours worth of audio content, which effectively translates into ~150k tokens, for tasks likes summarization and question answer.