Skip to content

[GLA-2952] Update audio_to_llm and summarization sections#117

Merged
remilejeune2 merged 8 commits into
mainfrom
chore/audio-to-llm
Apr 27, 2026
Merged

[GLA-2952] Update audio_to_llm and summarization sections#117
remilejeune2 merged 8 commits into
mainfrom
chore/audio-to-llm

Conversation

@remilejeune2
Copy link
Copy Markdown
Contributor

@remilejeune2 remilejeune2 commented Apr 20, 2026

Summary by CodeRabbit

  • Documentation
    • Rewrote audio-to-LLM guide into full user documentation: clarified execution flow (prompts run after transcription, one LLM response per prompt), request configuration options, model default/override behavior, updated examples including a post-meeting multi-prompt workflow, new top-level results structure, response shape, and pricing.
    • Updated summarization docs to specify default model, override field in config, example updates, and detailed pricing.

@coderabbitai
Copy link
Copy Markdown
Contributor

coderabbitai Bot commented Apr 20, 2026

Warning

Rate limit exceeded

@remilejeune2 has exceeded the limit for the number of commits that can be reviewed per hour. Please wait 54 minutes and 51 seconds before requesting another review.

To keep reviews running without waiting, you can enable usage-based add-on for your organization. This allows additional reviews beyond the hourly cap. Account admins can enable it under billing.

⌛ How to resolve this issue?

After the wait time has elapsed, a review can be triggered using the @coderabbitai review command as a PR comment. Alternatively, push new commits to this PR.

We recommend that you space out your commits to avoid hitting the rate limit.

🚦 How do rate limits work?

CodeRabbit enforces hourly rate limits for each developer per organization.

Our paid plans have higher rate limits than the trial, open-source and free plans. In all cases, we re-allow further reviews after a brief timeout.

Please see our FAQ for further information.

ℹ️ Review info
⚙️ Run configuration

Configuration used: defaults

Review profile: CHILL

Plan: Pro

Run ID: 573ce097-d642-4ca9-927b-2a47afc72509

📥 Commits

Reviewing files that changed from the base of the PR and between 9205201 and 3b8ed40.

📒 Files selected for processing (1)
  • chapters/audio-intelligence/summarization.mdx
📝 Walkthrough

Walkthrough

Rewrote audio-to-llm docs into a full product guide (execution flow, request/response shape, prompts, model selection, examples, pricing). Updated summarization docs to set default model, allow model override, update examples, and add pricing/context window details.

Changes

Cohort / File(s) Summary
Audio-to-LLM docs
chapters/audio-intelligence/audio-to-llm.mdx
Full rewrite from alpha note to a complete guide: explains that prompts run after transcription using transcript text only; adds audio_to_llm: true and audio_to_llm_config.prompts usage, per-prompt execution and outputs, optional model key (default GPT-5.4-nano / OpenRouter override), updated examples (single-prompt and post-meeting multi-prompt), new top-level results array response shape, and pricing section with token-cost tables.
Summarization docs
chapters/audio-intelligence/summarization.mdx
Adds default model openai/gpt-5.4-nano and instructions to override via summarization_config.model; updates example requests for pre-recorded and live flows to include model, and introduces pricing and context-window table for supported models.

Sequence Diagram(s)

sequenceDiagram
    participant Client as Client
    participant Ingest as Audio Ingest / Transcription
    participant Orchestrator as Audio-to-LLM Orchestrator
    participant LLM as LLM (OpenRouter / GPT-5.4-nano)

    Client->>Ingest: Upload audio / start live stream
    Ingest->>Orchestrator: Transcription (transcript text)
    Orchestrator->>Orchestrator: For each prompt in audio_to_llm_config.prompts (in-order)
    Orchestrator->>LLM: Send prompt + transcript (model per-config or default)
    LLM-->>Orchestrator: Response (per-prompt)
    Orchestrator-->>Client: Aggregate `results` array with per-prompt outputs
Loading

Estimated code review effort

🎯 2 (Simple) | ⏱️ ~10 minutes

Suggested reviewers

  • mfernandez-gladia

Poem

🐰 I hopped through words and lines anew,

Prompts and transcripts in tidy queue,
Models picked, the summaries sing,
Tokens counted — what a spring!
A rabbit cheers and nibbles glue. 🥕

🚥 Pre-merge checks | ✅ 5
✅ Passed checks (5 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title check ✅ Passed The title directly references the two main files/sections updated: audio_to_llm and summarization documentation with comprehensive rewrites and new model/pricing sections.
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.
Linked Issues check ✅ Passed Check skipped because no linked issues were found for this pull request.
Out of Scope Changes check ✅ Passed Check skipped because no linked issues were found for this pull request.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.

✨ Finishing Touches
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Commit unit tests in branch chore/audio-to-llm

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

@mintlify
Copy link
Copy Markdown
Contributor

mintlify Bot commented Apr 20, 2026

Preview deployment for your docs. Learn more about Mintlify Previews.

Project Status Preview Updated (UTC)
gladia 🟢 Ready View Preview Apr 20, 2026, 7:50 PM

💡 Tip: Enable Workflows to automatically generate PRs for you.

Copy link
Copy Markdown
Contributor

@coderabbitai coderabbitai Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 1

Caution

Some comments are outside the diff and can’t be posted inline due to platform limitations.

⚠️ Outside diff range comments (1)
chapters/audio-intelligence/summarization.mdx (1)

38-62: ⚠️ Potential issue | 🟡 Minor

Invalid JSON: trailing comma after "model" in both examples.

Lines 44 and 54 end the last property with a trailing comma, which is not valid JSON. Users copy-pasting these snippets into a strict JSON parser will get a syntax error.

🔧 Proposed fix
   "summarization_config": {
     "type": "concise",
-    "model": "openai/gpt-5.4-nano",
+    "model": "openai/gpt-5.4-nano"
   }
     "summarization_config": {
       "type": "concise",
-      "model": "openai/gpt-5.4-nano",
+      "model": "openai/gpt-5.4-nano"
     }
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@chapters/audio-intelligence/summarization.mdx` around lines 38 - 62, The JSON
examples inside the CodeGroup contain invalid trailing commas after the "model"
property in both the Pre-recorded block under "summarization_config" and the
Live block under "post_processing" -> "summarization_config"; remove the
trailing commas after "openai/gpt-5.4-nano" in both places so the
"summarization_config" objects are valid JSON (verify the surrounding braces
remain correct and no other trailing commas exist).
🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Inline comments:
In `@chapters/audio-intelligence/summarization.mdx`:
- Around line 88-94: Update the "Anthropic: Claude Opus 4.7" table row so the
pricing matches OpenRouter's listing (input $5/M, output $25/M) instead of the
current $19.50/$97.50; locate the row by the model label "Anthropic: Claude Opus
4.7" or config token `anthropic/claude-opus-4.7` and change the Input and Output
columns accordingly in both files where the row appears (summarization.mdx and
audio-to-llm.mdx).

---

Outside diff comments:
In `@chapters/audio-intelligence/summarization.mdx`:
- Around line 38-62: The JSON examples inside the CodeGroup contain invalid
trailing commas after the "model" property in both the Pre-recorded block under
"summarization_config" and the Live block under "post_processing" ->
"summarization_config"; remove the trailing commas after "openai/gpt-5.4-nano"
in both places so the "summarization_config" objects are valid JSON (verify the
surrounding braces remain correct and no other trailing commas exist).
🪄 Autofix (Beta)

Fix all unresolved CodeRabbit comments on this PR:

  • Push a commit to this branch (recommended)
  • Create a new PR with the fixes

ℹ️ Review info
⚙️ Run configuration

Configuration used: defaults

Review profile: CHILL

Plan: Pro

Run ID: 2b87911a-bfef-4fa6-907d-c9ea39c2d757

📥 Commits

Reviewing files that changed from the base of the PR and between fbd61b0 and bdb0f85.

📒 Files selected for processing (2)
  • chapters/audio-intelligence/audio-to-llm.mdx
  • chapters/audio-intelligence/summarization.mdx

Comment thread chapters/audio-intelligence/summarization.mdx Outdated
Copy link
Copy Markdown
Contributor

@coderabbitai coderabbitai Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 1

Caution

Some comments are outside the diff and can’t be posted inline due to platform limitations.

⚠️ Outside diff range comments (1)
chapters/audio-intelligence/summarization.mdx (1)

39-61: ⚠️ Potential issue | 🟡 Minor

Invalid JSON: trailing comma after "model" value.

Both the Pre-recorded and Live examples end the summarization_config object with a trailing comma ("model": "openai/gpt-5.4", before the closing }), which is invalid JSON and will fail for users who copy-paste these snippets into a strict JSON parser.

🔧 Proposed fix
   "summarization_config": {
     "type": "concise",
-    "model": "openai/gpt-5.4",
+    "model": "openai/gpt-5.4"
   }
     "summarization_config": {
       "type": "concise",
-      "model": "openai/gpt-5.4",
+      "model": "openai/gpt-5.4"
     }
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@chapters/audio-intelligence/summarization.mdx` around lines 39 - 61, Remove
the trailing commas after the "model" entries in both code examples so the JSON
becomes valid: update the Pre-recorded snippet's summarization_config and the
Live snippet's post_processing.summarization_config by deleting the comma after
"openai/gpt-5.4" and ensure each object closes cleanly (no dangling commas).
🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Inline comments:
In `@chapters/audio-intelligence/audio-to-llm.mdx`:
- Line 137: Update the pricing intro sentence "The input provided to the LLM is
the full transcription" in chapters/audio-intelligence/audio-to-llm.mdx to
clarify that audio-to-LLM billing includes the full transcript plus each prompt
sent to the model, and that prompts are billed once per prompt; explicitly note
that costs scale with number of prompts so users can estimate per-1M-token
pricing correctly.

---

Outside diff comments:
In `@chapters/audio-intelligence/summarization.mdx`:
- Around line 39-61: Remove the trailing commas after the "model" entries in
both code examples so the JSON becomes valid: update the Pre-recorded snippet's
summarization_config and the Live snippet's post_processing.summarization_config
by deleting the comma after "openai/gpt-5.4" and ensure each object closes
cleanly (no dangling commas).
🪄 Autofix (Beta)

Fix all unresolved CodeRabbit comments on this PR:

  • Push a commit to this branch (recommended)
  • Create a new PR with the fixes

ℹ️ Review info
⚙️ Run configuration

Configuration used: defaults

Review profile: CHILL

Plan: Pro

Run ID: b340751e-0889-4711-850f-c4fb83fa2f72

📥 Commits

Reviewing files that changed from the base of the PR and between bdb0f85 and ff48671.

📒 Files selected for processing (2)
  • chapters/audio-intelligence/audio-to-llm.mdx
  • chapters/audio-intelligence/summarization.mdx

Comment thread chapters/audio-intelligence/audio-to-llm.mdx Outdated
@Karamouche
Copy link
Copy Markdown
Contributor

A few feedbacks:

  • The "gladia fee" is a bit unclear : is that 10% more that openrouter or something like that? I think we need to be transparent on this
  • So those two features (summarization and audio-to-llm) will now have an extra cost, even with using gpt 5.4 nano?

Copy link
Copy Markdown
Contributor

@coderabbitai coderabbitai Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 1

Caution

Some comments are outside the diff and can’t be posted inline due to platform limitations.

⚠️ Outside diff range comments (1)
chapters/audio-intelligence/summarization.mdx (1)

39-61: ⚠️ Potential issue | 🔴 Critical

Invalid JSON: trailing comma after the new "model" field in both examples.

Lines 44 and 54 add "model": "openai/gpt-5.4", immediately before the closing }. Standard JSON does not allow trailing commas, so users who copy-paste these snippets verbatim into a JSON-validating client will get parse errors.

Note: the model identifier openai/gpt-5.4 here also doesn't match the documented default openai/gpt-5.4-nano — confirm this is intentional (showing an override) or align both.

🔧 Proposed fix
 ```json Pre-recorded
 {
   "summarization": true,
   "summarization_config": {
     "type": "concise",
-    "model": "openai/gpt-5.4",
+    "model": "openai/gpt-5.4"
   }
 }
{
  "post_processing": {
    "summarization": true,
    "summarization_config": {
      "type": "concise",
-      "model": "openai/gpt-5.4",
+      "model": "openai/gpt-5.4"
    }
  },
  "messages_config": {
    "receive_post_processing_events": true
  }
}
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@chapters/audio-intelligence/summarization.mdx` around lines 39 - 61, Remove
the trailing commas after the "model" entries in both JSON examples
("Pre-recorded" and "Live") so the snippets are valid JSON, and confirm whether
the "model" value should be "openai/gpt-5.4" or the documented default
"openai/gpt-5.4-nano" and make them consistent (either keep the explicit
override in both snippets or change both to the default model identifier).
♻️ Duplicate comments (1)
chapters/audio-intelligence/audio-to-llm.mdx (1)

135-137: ⚠️ Potential issue | 🟠 Major

Pricing intro is still understated for audio-to-LLM (per-prompt billing not conveyed).

Audio-to-LLM sends the transcript plus each prompt, executed once per prompt, so token usage scales linearly with the number of prompts. The current sentence reads identically to the summarization page and will lead users to underestimate costs (and ties into the reviewer's PR question about extra cost when using gpt-5.4-nano).

🔧 Proposed fix
-The input provided to the LLM is the full transcription. All prices are per 1M tokens and include platform fees (30%).
+Each prompt is executed independently with the full transcript plus the prompt as input, so token usage scales with the number of prompts. All prices are per 1M tokens and include platform fees (30%).
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@chapters/audio-intelligence/audio-to-llm.mdx` around lines 135 - 137, Update
the "## Pricing" paragraph to explicitly state that audio-to-LLM billing is per
prompt: the LLM receives the full transcription plus each prompt, and each
prompt is executed once per submission so token usage (and cost) scales linearly
with the number of prompts; keep the existing note that prices are per 1M tokens
and include platform fees (30%) but add a short example or clarifying clause
that multiple prompts multiply token usage to prevent underestimation.
🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Inline comments:
In `@chapters/audio-intelligence/summarization.mdx`:
- Around line 84-86: The Pricing section in summarization.mdx currently says
"include platform fees" but doesn't state the percentage; update the Pricing
paragraph to explicitly state the platform fee percentage (30%) so it matches
audio-to-llm.mdx and allows users to reproduce Gladia fee calculations from
OpenRouter rates; edit the "## Pricing" block text to append or replace the
sentence with a clear statement like "All prices are per 1M tokens and include a
30% platform fee (Gladia) — this is applied on top of base OpenRouter rates" so
the page is consistent and transparent.

---

Outside diff comments:
In `@chapters/audio-intelligence/summarization.mdx`:
- Around line 39-61: Remove the trailing commas after the "model" entries in
both JSON examples ("Pre-recorded" and "Live") so the snippets are valid JSON,
and confirm whether the "model" value should be "openai/gpt-5.4" or the
documented default "openai/gpt-5.4-nano" and make them consistent (either keep
the explicit override in both snippets or change both to the default model
identifier).

---

Duplicate comments:
In `@chapters/audio-intelligence/audio-to-llm.mdx`:
- Around line 135-137: Update the "## Pricing" paragraph to explicitly state
that audio-to-LLM billing is per prompt: the LLM receives the full transcription
plus each prompt, and each prompt is executed once per submission so token usage
(and cost) scales linearly with the number of prompts; keep the existing note
that prices are per 1M tokens and include platform fees (30%) but add a short
example or clarifying clause that multiple prompts multiply token usage to
prevent underestimation.
🪄 Autofix (Beta)

Fix all unresolved CodeRabbit comments on this PR:

  • Push a commit to this branch (recommended)
  • Create a new PR with the fixes

ℹ️ Review info
⚙️ Run configuration

Configuration used: defaults

Review profile: CHILL

Plan: Pro

Run ID: 2beacf11-3268-4309-badc-0e0d48d54799

📥 Commits

Reviewing files that changed from the base of the PR and between ff48671 and 9205201.

📒 Files selected for processing (2)
  • chapters/audio-intelligence/audio-to-llm.mdx
  • chapters/audio-intelligence/summarization.mdx

Comment thread chapters/audio-intelligence/summarization.mdx Outdated
@remilejeune2
Copy link
Copy Markdown
Contributor Author

A few feedbacks:

* The "gladia fee" is a bit unclear : is that 10% more that openrouter or something like that? I think we need to be transparent on this

Fixed.

* So those two features (summarization and audio-to-llm) will now have an extra cost, even with using gpt 5.4 nano?

Yes, default model will also be charged.

@remilejeune2 remilejeune2 merged commit 31c5f5f into main Apr 27, 2026
8 checks passed
@remilejeune2 remilejeune2 deleted the chore/audio-to-llm branch April 27, 2026 13:28
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants