gsaon commited on
Commit
f12c615
·
verified ·
1 Parent(s): 15571a0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -43,7 +43,7 @@ We evaluated granite-speech-4.1-2b alongside other speech-language models in the
43
 
44
  We evaluated the model’s keyword list biasing (KWB) capability by comparing performance with and without KWB applied at inference time.
45
  We report the F1 scores of transcribed keywords during ASR tasks, excluding common words from the evaluation.
46
- ![kwb-f1](https://cdn-uploads.huggingface.co/production/uploads/666ec38102791b3b49f453e8/4JkeIsHOy6BFL5DCnGVX6.png)
47
 
48
  We also evaluated our model on a variety of corpora to assess its punctuation and capitalization capabilities. We report the metrics as defined in [LibriSpeech-PC](https://arxiv.org/abs/2310.02943). PER (punctuation error rate) measures errors in the insertion, deletion, or substitution of punctuation marks (periods, commas, and question marks). Cap-F1 (capitalization F1) measures how accurately the model capitalizes relevant words in the output. Note that our Cap-F1 is computed on Levenshtein-aligned matching word pairs rather than fully matching sentences, allowing evaluation even in the presence of ASR errors.
49
 
 
43
 
44
  We evaluated the model’s keyword list biasing (KWB) capability by comparing performance with and without KWB applied at inference time.
45
  We report the F1 scores of transcribed keywords during ASR tasks, excluding common words from the evaluation.
46
+ ![kwb-f1.v2](https://cdn-uploads.huggingface.co/production/uploads/666ec38102791b3b49f453e8/nNb161Sc_DKC3TONbkbGK.jpeg)
47
 
48
  We also evaluated our model on a variety of corpora to assess its punctuation and capitalization capabilities. We report the metrics as defined in [LibriSpeech-PC](https://arxiv.org/abs/2310.02943). PER (punctuation error rate) measures errors in the insertion, deletion, or substitution of punctuation marks (periods, commas, and question marks). Cap-F1 (capitalization F1) measures how accurately the model capitalizes relevant words in the output. Note that our Cap-F1 is computed on Levenshtein-aligned matching word pairs rather than fully matching sentences, allowing evaluation even in the presence of ASR errors.
49