Speech & Audio¶
This section details how to evaluate speech and audio benchmarks, including understanding tasks that test models' ability to reason about audio content (speech, music, environmental sounds) and ASR tasks for transcription.
Running without audio files
If you want to evaluation without audio files (not recommended) use
--no-audio flag. In this case you can also set --skip_data_dir_check
as data is very lightweight when audio files aren't being used.
Supported benchmarks¶
ASR Leaderboard¶
ASR benchmark based on the HuggingFace Open ASR Leaderboard. Evaluates transcription quality using Word Error Rate (WER).
Datasets: librispeech_clean, librispeech_other, voxpopuli, tedlium, gigaspeech, spgispeech, earnings22, ami
Dataset Location¶
- Benchmark is defined in
nemo_skills/dataset/asr-leaderboard/__init__.py - Original datasets are hosted on HuggingFace (downloaded automatically during preparation)
MMAU-Pro¶
MMAU-Pro (Multimodal Audio Understanding - Pro) is a comprehensive benchmark for evaluating audio understanding capabilities across three different task categories:
- Closed-form questions: Questions with specific answers evaluated using NVEmbed similarity matching
- Open-ended questions: Questions requiring detailed responses, evaluated with LLM-as-a-judge (Qwen 2.5)
- Instruction following: Tasks that test the model's ability to follow audio-related instructions
Dataset Location¶
- Benchmark is defined in
nemo_skills/dataset/mmau-pro/__init__.py - Original benchmark source is hosted on HuggingFace
Preparing Data¶
These benchmarks require audio files for meaningful evaluation. Audio files are downloaded by default to ensure proper evaluation.
Data Preparation¶
To prepare the dataset with audio files:
Prepare specific datasets only:
MMAU-Pro¶
Running Evaluation¶
ASR Leaderboard¶
from nemo_skills.pipeline.cli import wrap_arguments, eval
eval(
ctx=wrap_arguments(""),
cluster="oci_iad",
output_dir="/workspace/asr-leaderboard-eval",
benchmarks="asr-leaderboard",
server_type="megatron",
server_gpus=1,
model="/workspace/checkpoint",
server_entrypoint="/workspace/megatron-lm/server.py",
server_container="/path/to/container.sqsh",
data_dir="/dataset",
installation_command="pip install -r requirements/audio.txt",
server_args="--inference-max-requests 1 --model-config /workspace/checkpoint/config.yaml",
)
Evaluate a specific dataset:
Alternative: Command-line usage
ns eval \
--cluster=oci_iad \
--output_dir=/workspace/path/to/asr-leaderboard-eval \
--benchmarks=asr-leaderboard \
--server_type=megatron \
--server_gpus=1 \
--model=/workspace/path/to/checkpoint \
--server_entrypoint=/workspace/megatron-lm/server.py \
--server_container=/path/to/container.sqsh \
--data_dir=/dataset \
--installation_command="pip install -r requirements/audio.txt"
MMAU-Pro¶
import os
from nemo_skills.pipeline.cli import wrap_arguments, eval
os.environ["NVIDIA_API_KEY"] = "your_nvidia_api_key" # For LLM judge
eval(
ctx=wrap_arguments(""),
cluster="oci_iad",
output_dir="/workspace/mmau-pro-eval",
benchmarks="mmau-pro",
server_type="megatron",
server_gpus=1,
model="/workspace/checkpoint",
server_entrypoint="/workspace/megatron-lm/server.py",
server_container="/path/to/container.sqsh",
data_dir="/dataset",
installation_command="pip install sacrebleu",
server_args="--inference-max-requests 1 --model-config /workspace/checkpoint/config.yaml",
)
Evaluate individual categories:
mmau-pro.closed_formmmau-pro.open_endedmmau-pro.instruction_following
Alternative: Command-line usage
export NVIDIA_API_KEY=your_nvidia_api_key
ns eval \
--cluster=oci_iad \
--output_dir=/workspace/path/to/mmau-pro-eval \
--benchmarks=mmau-pro \
--server_type=megatron \
--server_gpus=1 \
--model=/workspace/path/to/checkpoint \
--server_entrypoint=/workspace/megatron-lm/server.py \
--server_container=/path/to/container.sqsh \
--data_dir=/dataset \
--installation_command="pip install sacrebleu"
Using Custom Judge Models¶
The open-ended questions subset uses an LLM-as-a-judge (by default, Qwen 2.5 7B via NVIDIA API) to evaluate responses. You can customize the judge model for this subset:
from nemo_skills.pipeline.cli import wrap_arguments, eval
import os
os.environ["NVIDIA_API_KEY"] = "your_nvidia_api_key"
eval(
ctx=wrap_arguments(""),
cluster="oci_iad",
output_dir="/workspace/path/to/mmau-pro-eval",
benchmarks="mmau-pro.open_ended", # Only open-ended uses LLM judge
server_type="megatron",
server_gpus=1,
model="/workspace/path/to/checkpoint-tp1",
# ... other server args ...
)
Self-hosted Judge Limitation
When using a self-hosted judge, evaluate mmau-pro.open_ended separately.
from nemo_skills.pipeline.cli import wrap_arguments, eval
eval(
ctx=wrap_arguments("++prompt_suffix='/no_think'"),
cluster="oci_iad",
output_dir="/workspace/path/to/mmau-pro-eval",
benchmarks="mmau-pro.open_ended", # Only open-ended uses LLM judge
server_type="megatron",
server_gpus=1,
model="/workspace/path/to/checkpoint-tp1",
# Judge configuration
judge_model="Qwen/Qwen2.5-32B-Instruct",
judge_server_type="sglang",
judge_server_gpus=2,
# ... other server args ...
)
Understanding Results¶
After evaluation completes, results are saved in your output directory under eval-results/.
ASR Leaderboard Results¶
Example output:
------------------------------------- asr-leaderboard --------------------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | wer | num_entries
pass@1 | 736 | 233522 | 86.70% | 0.00% | 7.82% | 143597
----------------------------------- asr-leaderboard-ami ------------------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | wer | num_entries
pass@1 | 732 | 3680 | 81.27% | 0.00% | 18.45% | 12620
-------------------------------- asr-leaderboard-earnings22 --------------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | wer | num_entries
pass@1 | 736 | 3522 | 83.97% | 0.00% | 14.72% | 57390
-------------------------------- asr-leaderboard-gigaspeech --------------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | wer | num_entries
pass@1 | 736 | 233469 | 71.86% | 0.00% | 12.34% | 25376
---------------------------- asr-leaderboard-librispeech_clean ----------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | wer | num_entries
pass@1 | 735 | 3607 | 99.62% | 0.00% | 2.06% | 2620
---------------------------- asr-leaderboard-librispeech_other ----------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | wer | num_entries
pass@1 | 733 | 3927 | 98.67% | 0.00% | 4.34% | 2939
-------------------------------- asr-leaderboard-spgispeech -------------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | wer | num_entries
pass@1 | 740 | 4510 | 99.99% | 0.00% | 3.81% | 39341
--------------------------------- asr-leaderboard-tedlium ----------------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | wer | num_entries
pass@1 | 732 | 3878 | 77.74% | 0.00% | 7.89% | 1469
-------------------------------- asr-leaderboard-voxpopuli --------------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | wer | num_entries
pass@1 | 741 | 4007 | 99.51% | 0.00% | 6.47% | 1842
MMAU-Pro Results¶
<output_dir>/
├── eval-results/
│ └── mmau-pro/
│ ├── metrics.json # Overall aggregate scores
│ ├── mmau-pro.instruction_following/
│ │ └── metrics.json
│ ├── mmau-pro.closed_form/
│ │ └── metrics.json
│ └── mmau-pro.open_ended/
│ └── metrics.json
Example output:
Open-Ended Questions:
------------------------------- mmau-pro.open_ended -------------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 82 | 196 | 14.88% | 0.00% | 625
Instruction Following:
-------------------------- mmau-pro.instruction_following -------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 0 | 102 | 21.84% | 0.00% | 87
Closed-Form Questions (Main Category + Sub-categories):
------------------------------- mmau-pro.closed_form ------------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 2 | 6581 | 33.88% | 0.00% | 4593
---------------------------- mmau-pro.closed_form-sound ---------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 0 | 691 | 26.15% | 0.00% | 1048
---------------------------- mmau-pro.closed_form-multi ---------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 1 | 6005 | 24.65% | 0.00% | 430
------------------------- mmau-pro.closed_form-sound_music ------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 0 | 810 | 22.00% | 0.00% | 50
---------------------------- mmau-pro.closed_form-music ---------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 5 | 5467 | 42.81% | 0.00% | 1418
------------------------ mmau-pro.closed_form-spatial_audio -----------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 0 | 5597 | 2.15% | 0.00% | 325
------------------------ mmau-pro.closed_form-music_speech ------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 1 | 5658 | 36.96% | 0.00% | 46
--------------------- mmau-pro.closed_form-sound_music_speech ---------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 0 | 5664 | 14.29% | 0.00% | 7
------------------------ mmau-pro.closed_form-sound_speech ------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 0 | 5713 | 36.36% | 0.00% | 88
--------------------------- mmau-pro.closed_form-speech ---------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 1 | 6312 | 38.16% | 0.00% | 891
------------------------- mmau-pro.closed_form-voice_chat -------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 0 | 6580 | 55.52% | 0.00% | 290
Overall Aggregate Score:
-------------------------------- mmau-pro -----------------------------------------
evaluation_mode | avg_tokens | gen_seconds | success_rate | no_answer | num_entries
pass@1 | 11 | 6879 | 31.44% | 0.00% | 5305
AudioBench¶
AudioBench is a comprehensive benchmark for evaluating speech and audio language models across multiple tasks including ASR, translation, speech QA, and audio understanding.
Dataset Location¶
- Benchmark is defined in
nemo_skills/dataset/audiobench/__init__.py - External source repository is AudioBench
Data Preparation¶
AudioBench can be prepared via the NeMo-Skills data preparation entrypoint. By default it will download/copy audio files into the prepared dataset directory.
To prepare without saving audio files (not recommended):
LibriSpeech-PC¶
LibriSpeech-PC is an Automatic Speech Recognition (ASR) benchmark that evaluates models' ability to transcribe speech with proper punctuation and capitalization. It builds upon the original LibriSpeech corpus with enhanced reference transcripts.
Dataset Location¶
- Benchmark is defined in
nemo_skills/dataset/librispeech-pc/__init__.py - Manifests (with punctuation/capitalization) from OpenSLR-145
- Audio files from original LibriSpeech OpenSLR-12
Available Splits¶
test-clean: Clean speech recordings (easier subset)test-other: More challenging recordings with varied acoustic conditions
Preparing LibriSpeech-PC Data¶
LibriSpeech-PC requires audio files for ASR evaluation. Audio files are downloaded by default.
Data Preparation¶
To prepare the dataset with audio files:
Preparing Specific Splits¶
To prepare only one split:
or
Numb3rs¶
Numb3rs is a speech benchmark for evaluating text normalization (TN) and inverse text normalization (ITN) capabilities of audio-language models. It contains paired written/spoken forms with corresponding synthetic audio, allowing evaluation of whether a model transcribes numbers in written form (e.g., $100, 3.14) or spoken form (e.g., one hundred dollars, three point one four).
Dataset: nvidia/Numb3rs on HuggingFace
Categories: ADDRESS, CARDINAL, DATE, DECIMAL, DIGIT, FRACTION, MEASURE, MONEY, ORDINAL, PLAIN, TELEPHONE, TIME
Size: ~10K samples, ~4.89h total audio duration
Dataset Location¶
- Benchmark is defined in
nemo_skills/dataset/numb3rs/__init__.py - Original dataset is hosted on HuggingFace
Key Features¶
- Dual reference evaluation: Each sample has both a written form (
text_tn) and a spoken form (text_itn). WER is computed against both references. - Three prompt variants generated as separate split files:
test_neutral: Neutral transcription prompt ("Transcribe the audio file into English text.")test_tn: Text normalization prompt — expects written form (e.g.,$100)test_itn: Inverse text normalization prompt — expects spoken form (e.g.,one hundred dollars)
- Special normalization mode
no_tn_itn: Applies only lowercase + punctuation removal (no whisper normalization that would convert number words to digits, which would defeat the purpose of TN/ITN evaluation).
Preparing Numb3rs Data¶
Numb3rs requires audio files for evaluation. Audio files are downloaded by default from HuggingFace.
To prepare without saving audio files:
Prepare specific categories only:
Set a custom audio path prefix (for non-standard mount points):
Running Numb3rs Evaluation¶
The --split flag selects the prompt variant:
# Neutral prompt (default)
ns eval --benchmarks=numb3rs:1 --split=test_neutral ...
# Text normalization prompt (expects written form, e.g. "$100")
ns eval --benchmarks=numb3rs:1 --split=test_tn ...
# Inverse text normalization prompt (expects spoken form, e.g. "one hundred dollars")
ns eval --benchmarks=numb3rs:1 --split=test_itn ...
Understanding Numb3rs Results¶
Numb3rs reports the following metrics:
- wer: Word Error Rate against the expected answer (written form for TN, spoken form for ITN/neutral)
- wer_tn: WER against the written form reference (
text_tn) - wer_itn: WER against the spoken form reference (
text_itn) - success_rate: Percentage of samples with WER < 0.5
Per-category breakdowns (e.g., numb3rs-numb3rs_CARDINAL, numb3rs-numb3rs_MONEY) are included automatically.