# Evaluating SeamlessM4T models Refer to the [SeamlessM4T README](../../../../../docs/m4t) for an overview of the M4T models. Refer to the [inference README](../predict/README.md) for how to run inference with SeamlessM4T models. ## Quick start: We use SACREBLEU library for computing BLEU scores and [JiWER library](https://github.com/jitsi/jiwer) is used to compute these CER and WER scores. Evaluation can be run with the CLI, from the root directory of the repository. The model can be specified with `--model_name`: `seamlessM4T_v2_large` or `seamlessM4T_large` or `seamlessM4T_medium` ```bash m4t_evaluate --data_file --task --tgt_lang --output_path --ref_field --audio_root_dir ``` ## Note 1. We use raw (unnormalized) references to compute BLEU scores for S2TT, T2TT tasks. 2. For ASR task, src_lang needs to be passed as 3. `--src_lang` arg needs to be specified to run evaluation for T2TT task