Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
28 commits
Select commit Hold shift + click to select a range
eb98654
Debug logging for evaluate_generated_audio.py
rfejgin Feb 7, 2026
decc8c9
Move codec codes to device (NeMo Skills compatibility)
rfejgin Feb 7, 2026
0ed3c8b
Fix UTMOSv2 model caching
rfejgin Feb 7, 2026
52d2d39
Debug logging for FCD metric
rfejgin Feb 7, 2026
bb58546
Fix logic on when to update FCD metric
rfejgin Feb 7, 2026
adf7d6b
Tweak UTMOSv2 cache location setup
rfejgin Feb 9, 2026
85cdc04
Fix phoneme_dict and heteronyms paths in MagpieTTS model loading
rfejgin Feb 10, 2026
8c0bc3c
Delete temporary debug logging
rfejgin Feb 10, 2026
ee06675
Log evaluation timing breakdown for profiling
rfejgin Feb 11, 2026
fc23ea2
Experimental: batch ASRs during evaluation
rfejgin Feb 11, 2026
ed5d829
Experimental: tune ASR batch size during evaluation
rfejgin Feb 11, 2026
80ae46b
Experimental: tune UTMOSv2 batch size during evaluation
rfejgin Feb 11, 2026
0494d62
Experimental: make evaluation compatible with chunked operation (used…
rfejgin Feb 12, 2026
e37aa88
Merge remote-tracking branch 'nemo/main' into nemo-main-rfejgin
rfejgin Feb 20, 2026
e66337b
Cleanup
rfejgin Feb 20, 2026
4589a8c
Cleanup utmos thread / batch size
rfejgin Feb 20, 2026
20f19f9
Add back filewise metric filtering
rfejgin Feb 21, 2026
85ff1a7
Remove non-batched version of whisper transcription
rfejgin Feb 21, 2026
8fc8190
Refactor evaluate_generated_audio.py
rfejgin Feb 21, 2026
7ae45be
Merge remote-tracking branch 'nemo/main' into magpietts_evaluation_pa…
rfejgin Feb 21, 2026
4a66481
Merge branch 'main' into magpietts_evaluation_parallelization
rfejgin Feb 23, 2026
23e86bc
Refactoring
rfejgin Feb 23, 2026
fd7b708
Cleanup
rfejgin Feb 23, 2026
68e4b1c
Merge branch 'magpietts_evaluation_parallelization' of github.com:rfe…
rfejgin Feb 23, 2026
38b0ea3
Cleanup
rfejgin Feb 23, 2026
a96c33d
Restore a logging warning that was accidentally removed.
rfejgin Feb 23, 2026
7340a13
Make device used in evaluation configurable.
rfejgin Feb 24, 2026
2dda783
Merge branch 'main' into magpietts_evaluation_parallelization
rfejgin Feb 24, 2026
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions examples/tts/magpietts_inference.py
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,6 @@
# Import the modular components
from nemo.collections.tts.modules.magpietts_inference.evaluation import (
DEFAULT_VIOLIN_METRICS,
STANDARD_METRIC_KEYS,
EvaluationConfig,
compute_mean_with_confidence_interval,
evaluate_generated_audio_dir,
Expand Down Expand Up @@ -292,6 +291,7 @@ def run_inference_and_evaluation(
with_utmosv2=eval_config.with_utmosv2,
with_fcd=eval_config.with_fcd,
codec_model_path=eval_config.codec_model_path,
device=eval_config.device,
)

metrics, filewise_metrics = evaluate_generated_audio_dir(
Expand All @@ -308,8 +308,9 @@ def run_inference_and_evaluation(
with open(os.path.join(eval_dir, f"{dataset}_metrics_{repeat_idx}.json"), "w") as f:
json.dump(metrics, f, indent=4)

sorted_filewise = sorted(filewise_metrics, key=lambda x: x.get('cer', 0), reverse=True)
with open(os.path.join(eval_dir, f"{dataset}_filewise_metrics_{repeat_idx}.json"), "w") as f:
json.dump(filewise_metrics, f, indent=4)
json.dump(sorted_filewise, f, indent=4)

# Append to per-run CSV
append_metrics_to_csv(per_run_csv, full_checkpoint_name, dataset, metrics)
Expand All @@ -331,7 +332,6 @@ def run_inference_and_evaluation(
# Compute mean with confidence interval across repeats
metrics_mean_ci = compute_mean_with_confidence_interval(
metrics_all_repeats,
STANDARD_METRIC_KEYS,
confidence=confidence_level,
)

Expand Down
2 changes: 0 additions & 2 deletions nemo/collections/tts/modules/magpietts_inference/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,6 @@

from nemo.collections.tts.modules.magpietts_inference.evaluation import (
DEFAULT_VIOLIN_METRICS,
STANDARD_METRIC_KEYS,
EvaluationConfig,
compute_mean_with_confidence_interval,
evaluate_generated_audio_dir,
Expand Down Expand Up @@ -74,7 +73,6 @@
"EvaluationConfig",
"evaluate_generated_audio_dir",
"compute_mean_with_confidence_interval",
"STANDARD_METRIC_KEYS",
"DEFAULT_VIOLIN_METRICS",
# Visualization
"create_violin_plot",
Expand Down
Loading
Loading