Announcements

New releases, model updates, awards, events, and more.

AssemblyAI ALD confidence threshold
Automatic language detection improvements: increased accuracy & expanded language support

Announcing several improvements to our automatic language detection model, yielding a more accurate model that's available in more languages.

AssemblyAI logo and Zapier logo
Introducing the enhanced AssemblyAI app for Zapier

The enhanced AssemblyAI app for Zapier is more flexible, supports more AssemblyAI features, and integrates more closely into the Zap editor.

Announcing New Language Support for PII Text Redaction and Expanding Entity Detection
Announcing New Language Support for PII Text Redaction and Expanding Entity Detection

We're announcing support for 47 new languages and 16 new entities to make our Audio Intelligence more powerful and globally accessible.

Claude 3 Models now available with LeMUR
Claude 3 Models now available with LeMUR

Today we're releasing Anthropic's Claude 3 model family into LeMUR as part of our ongoing commitment to giving you access to the most advanced and innovative AI capabilities available on the market.

Speaker diarization improvements: new languages, increased accuracy
Speaker diarization improvements: new languages, increased accuracy

Announcing several improvements to our Speaker Diarization service, yielding a more accurate model that's available in more languages.

Best and Nano Tiers: More Speech-to-Text and Pricing Options
Best and Nano Tiers: More Speech-to-Text and Pricing Options

Last month, we introduced the Best and Nano tiers — new pricing options for Speech-to-Text models to help you balance accuracy, cost, and speed. Here's a breakdown of our Best and Nano tiers.

AssemblyAI and Make logo
Introducing the AssemblyAI app for Make (Integromat)

We've been working with Make to make our Speech AI available to no-code builders. With the AssemblyAI app for Make, you can use our speech recognition models, audio intelligence models, and build generative features with LLMs.

Real-Time is now Streaming Speech-to-Text, with added customization and control for users
Real-Time is now Streaming Speech-to-Text, with added customization and control for users

Streaming Speech-to-Text makes it easier than ever to transcribe live audio and videos, now with customizable end-of-utterance detection at a lower cost.

A New Free Python Course to Build Real-World Audio AI Apps
A New Free Python Course to Build Real-World Audio AI Apps

Learn how to build a real-world application working with audio data in this free video course.

AssemblyAI Go SDK v1.3.0: Utterance Detection and Word Search
AssemblyAI Go SDK v1.3.0: Utterance Detection and Word Search

Learn about the improvements in the recently released v1.3.0 of the AssemblyAI Go SDK.

Announcing the AssemblyAI Go SDK
Announcing the AssemblyAI Go SDK

We're excited to announce the release of our AssemblyAI SDK for Go!

Announcing the AssemblyAI Integration for Haystack
Announcing the AssemblyAI Integration for Haystack

Transcribe, summarize, or extract speakers from your audio files using the new AssemblyAI Haystack integration.

Lower latency, lower cost, more possibilities
Lower latency, lower cost, more possibilities

We’re excited to introduce major improvements to our API’s inference latency, with the majority of audio files now completing in well under 45 seconds regardless of audio duration.

Java code to transcribe an audio file using the AssemblyAI Java SDK.
Introducing the AssemblyAI Java SDK

We are thrilled to release the AssemblyAI Java SDK. You can use the SDK to transcribe audio asynchronously or in real-time, use our audio intelligence model, and apply LLMs to your audio data using LeMUR.

AssemblyAI and Semantic Kernel logo
Announcing the AssemblyAI integration for Semantic Kernel .NET

You can now integrate spoken audio data into Semantic Kernel .NET applications using the new AssemblyAI integration.

JavaScript sample for transcribing an audio file using LlamaIndex.TS
Announcing the AssemblyAI integration for LlamaIndex.TS

Transcribe audio files in your LlamaIndex.TS applications using the new AssemblyAI speech-to-text integration.

Announcing our $50M Series C to build superhuman Speech AI models
Announcing our $50M Series C to build superhuman Speech AI models

We're excited to share that we’ve raised $50M in Series C funding led by Accel, our partners that also led our Series A, with participation from Keith Block and Smith Point Capital, Insight Partners, Daniel Gross and Nat Friedman, and Y Combinator.

AssemblyAI is now on the Amazon Web Services (AWS) Marketplace
AssemblyAI is now on the Amazon Web Services (AWS) Marketplace

AssemblyAI is now an official partner on the Amazon Web Services (AWS) Marketplace.

Introducing Our New Punctuation Restoration and Truecasing Models
Introducing Our New Punctuation Restoration and Truecasing Models

We’ve trained new Punctuation and Truecasing models on 13 billion words to achieve a 39% F1 score improvement for mixed-case words. Building on a novel application of a hybrid architecture for a character-level classifier reduces inference time and improves the scalability of our Speech AI systems.

TypeScript code to transcribe an audio file and log it to the console.
Announcing the AssemblyAI Node SDK 2.0

We're thrilled to release version 2.0 of the AssemblyAI Node SDK. The SDK has been rewritten from scratch using TypeScript and provides easy-to-use access to our API.

AssemblyAI logo and Rivet logo
Announcing the AssemblyAI plugin for Rivet

AssemblyAI transcription and LeMUR capabilities are part of Rivet, the IDE for developing AI applications.

Introducing the AssemblyAI integration for LangChain.js
Introducing the AssemblyAI integration for LangChain.js

You can now integrate spoken audio data into LangChain.js applications using the new AssemblyAI integration.

Introducing LeMUR
Introducing LeMUR

LeMUR is the easiest way to build LLM apps on spoken data - search, summarize, and ask questions, with knowledge of your spoken data.

Conformer-2: a state-of-the-art speech recognition model trained on 1.1M hours of data
Conformer-2: a state-of-the-art speech recognition model trained on 1.1M hours of data

We're introducing Conformer-2, our latest AI model for automatic speech recognition. Conformer-2 is trained on 1.1M hours of English audio data, extending Conformer-1 to provide improvements on proper nouns, alphanumerics, and robustness to noise.

Introducing LeMUR, our new framework for applying powerful LLMs to transcribed speech
Introducing LeMUR, our new framework for applying powerful LLMs to transcribed speech

LeMUR is our new framework for applying powerful LLMs to transcribed speech. With a single line of code, LeMUR can quickly process audio transcripts for up to 10 hours worth of audio content, which effectively translates into ~150k tokens, for tasks likes summarization and question answer.