Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion chapters/audio-intelligence/audio-to-llm.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@
## Usage

1. Include `audio_to_llm: true` and an `audio_to_llm_config` object (at minimum, a `prompts` array) in your [pre-recorded transcription request](/chapters/pre-recorded-stt/quickstart).
2. Gladia transcribes the audio, along with any other audio-intelligence options you enabled on that request.

Check warning on line 17 in chapters/audio-intelligence/audio-to-llm.mdx

View check run for this annotation

Mintlify / Mintlify Validation (gladia-95) - vale-spellcheck

chapters/audio-intelligence/audio-to-llm.mdx#L17

Did you really mean 'Gladia'?
3. Each prompt is run on the resulting transcript via the LLM.
4. The API returns **one result object per prompt** (same order as `prompts`), each containing the original `prompt` and the model `response`.

Expand All @@ -26,7 +26,7 @@

By default the model used to execute your prompts is **[GPT 5.4 Nano](https://openrouter.ai/openai/gpt-5.4-nano)** (`openai/gpt-5.4-nano`), a fast option suited to high-volume summaries and extraction. The `model` can be customized when you need stronger reasoning, richer analysis, longer outputs, or behavior that fits a specific model.

You can use **any model listed on [OpenRouter](https://openrouter.ai/models)** by setting the `model` key. Prices reflect the public OpenRouter rate plus a platform fee added by Gladia.

Check warning on line 29 in chapters/audio-intelligence/audio-to-llm.mdx

View check run for this annotation

Mintlify / Mintlify Validation (gladia-95) - vale-spellcheck

chapters/audio-intelligence/audio-to-llm.mdx#L29

Did you really mean 'Gladia'?

## Example

Expand Down Expand Up @@ -134,7 +134,7 @@

## Pricing

The input provided to the LLM is the full transcription. All prices are per 1M tokens and include platform fees (30%).
The input provided to the LLM is the full transcription. All prices are per 1M tokens and include platform fees.
Comment thread
remilejeune2 marked this conversation as resolved.

| Model | `model` config | Context Window | Input | Output |
|-------|----------------|----------------|----------------|-----------------|
Expand Down
26 changes: 3 additions & 23 deletions chapters/audio-intelligence/summarization.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@
import PrerecordedBadge from "/snippets/badges/prerecorded.mdx"
import LiveBadge from "/snippets/badges/live.mdx"


<PrerecordedBadge />
<LiveBadge />

Expand All @@ -25,23 +26,16 @@

- **general**: Balanced summary for most use cases; good readability and coverage.
- **concise**: Shorter output for quick overviews or previews; fewer details.
- **bullet_points**: Lists key takeaways; ideal for action items, meeting notes, or highlights.

Check warning on line 29 in chapters/audio-intelligence/summarization.mdx

View check run for this annotation

Mintlify / Mintlify Validation (gladia-95) - vale-spellcheck

chapters/audio-intelligence/summarization.mdx#L29

Did you really mean 'bullet_points'?

## Model selection

By default the model used for summarization is **[GPT 5.4 Nano](https://openrouter.ai/openai/gpt-5.4-nano)** (`openai/gpt-5.4-nano`), a fast option suited to high-volume summaries and extraction. Use the **`model`** field inside **`summarization_config`** to override the default when stronger reasoning, richer analysis, longer outputs, or another provider or model family is needed.

You can use **any model listed on [OpenRouter](https://openrouter.ai/models)** by setting the `model` key. Prices reflect the public OpenRouter rate plus a platform fee added by Gladia.

## Usage
To enable summarization simply set the `"summarization"` parameter to true
<CodeGroup>
```json Pre-recorded
{
"summarization": true,
"summarization_config": {
"type": "concise",
"model": "openai/gpt-5.4",
"type": "concise"
}
}
```
Expand All @@ -50,8 +44,7 @@
"post_processing": {
"summarization": true,
"summarization_config": {
"type": "concise",
"model": "openai/gpt-5.4",
"type": "concise"
}
},
"messages_config": {
Expand Down Expand Up @@ -80,16 +73,3 @@
</CodeGroup>

You'll find the summarization of your audio under the `results` key.

## Pricing

The input provided to the LLM is the full transcription. All prices are per 1M tokens and include platform fees (30%).

| Model | `model` config | Context Window | Input | Output |
|-------|----------------|----------------|----------------|-----------------|
| [OpenAI: GPT-5.4 Nano](https://openrouter.ai/openai/gpt-5.4-nano) | `openai/gpt-5.4-nano` | 400k | $0.26 | $1.76 |
| [OpenAI: GPT-5.4](https://openrouter.ai/openai/gpt-5.4) | `openai/gpt-5.4` | 1.1M | $3.25 | $19.50 |
| [Anthropic: Claude Opus 4.7](https://openrouter.ai/anthropic/claude-opus-4.7) | `anthropic/claude-opus-4.7` | 1M | $6.50 | $32.50 |
| [Google: Gemini 3.1 Pro Preview](https://openrouter.ai/google/gemini-3.1-pro-preview) | `google/gemini-3.1-pro-preview` | 1M | $2.60 | $15.60 |
| [xAI: Grok 4.20](https://openrouter.ai/x-ai/grok-4.20) | `x-ai/grok-4.20` | 2M | $2.60 | $7.80 |
| [Meta: Llama 4 Maverick](https://openrouter.ai/meta-llama/llama-4-maverick) | `meta-llama/llama-4-maverick` | 1M | $0.20 | $0.78 |
Loading