diff --git a/recipes/GigaSpeech/ASR/CTC/train_with_wavlm.py b/recipes/GigaSpeech/ASR/CTC/train_with_wavlm.py index d2c1dc8152..60e1f6bb4a 100644 --- a/recipes/GigaSpeech/ASR/CTC/train_with_wavlm.py +++ b/recipes/GigaSpeech/ASR/CTC/train_with_wavlm.py @@ -44,7 +44,7 @@ def compute_forward(self, batch, stage): # Forward pass - # Handling SpeechBrain vs HuggingFance pretrained models + # Handling SpeechBrain vs HuggingFace pretrained models if hasattr(self.modules, "extractor"): # SpeechBrain pretrained model latents = self.modules.extractor(wavs) feats = self.modules.encoder_wrapper(latents, wav_lens=wav_lens)[ diff --git a/recipes/LibriSpeech/ASR/CTC/train_with_wav2vec_k2.py b/recipes/LibriSpeech/ASR/CTC/train_with_wav2vec_k2.py index 3501a268fa..2ddc4b72be 100644 --- a/recipes/LibriSpeech/ASR/CTC/train_with_wav2vec_k2.py +++ b/recipes/LibriSpeech/ASR/CTC/train_with_wav2vec_k2.py @@ -55,7 +55,7 @@ def compute_forward(self, batch, stage): # Forward pass - # Handling SpeechBrain vs HuggingFance pretrained models + # Handling SpeechBrain vs HuggingFace pretrained models if hasattr(self.modules, "extractor"): # SpeechBrain pretrained model latents = self.modules.extractor(wavs) feats = self.modules.encoder_wrapper(latents, wav_lens=wav_lens)[