Flurin17 commited on
Commit
a085a59
1 Parent(s): a00aeef

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -10
README.md CHANGED
@@ -28,11 +28,22 @@ Trained on RTX 3070 for 30 hours using SwissDial all Dialects with following gui
28
 
29
  ## Uses
30
 
 
 
 
 
 
 
 
 
 
 
 
31
 
32
  from peft import PeftModel, PeftConfig
33
  from transformers import WhisperForConditionalGeneration, Seq2SeqTrainer
34
 
35
- peft_model_id = "Flurin17/whisper-large-v3-peft-swiss-german" # Use the same model ID as before.
36
  peft_config = PeftConfig.from_pretrained(peft_model_id)
37
  model = WhisperForConditionalGeneration.from_pretrained(
38
  peft_config.base_model_name_or_path, load_in_8bit=True, device_map="auto"
@@ -41,21 +52,14 @@ model = PeftModel.from_pretrained(model, peft_model_id)
41
  model.config.use_cache = True
42
 
43
 
44
- from transformers import WhisperFeatureExtractor
45
-
46
- feature_extractor = WhisperFeatureExtractor.from_pretrained(peft_model_id)
47
-
48
- from transformers import WhisperTokenizer
49
-
50
- tokenizer = WhisperTokenizer.from_pretrained(peft_model_id, language=language, task=task)
51
-
52
  from transformers import AutomaticSpeechRecognitionPipeline
53
  import torch
54
  pipe = AutomaticSpeechRecognitionPipeline(model=model, tokenizer=tokenizer, feature_extractor=feature_extractor)
55
 
56
  with torch.cuda.amp.autocast():
57
- result = pipe(r"L:\Coding\random\audio.mp3", generate_kwargs={"language": "german"})
58
  print(result["text"])
 
59
 
60
 
61
 
 
28
 
29
  ## Uses
30
 
31
+ ```
32
+ model_name_or_path = "openai/whisper-large-v3"
33
+ task = "transcribe"
34
+ import json
35
+ import os
36
+ from transformers import WhisperFeatureExtractor
37
+ from transformers import WhisperTokenizer
38
+
39
+ feature_extractor = WhisperFeatureExtractor.from_pretrained(model_name_or_path)
40
+ tokenizer = WhisperTokenizer.from_pretrained(model_name_or_path, task=task)
41
+
42
 
43
  from peft import PeftModel, PeftConfig
44
  from transformers import WhisperForConditionalGeneration, Seq2SeqTrainer
45
 
46
+ peft_model_id = "flurin17/whisper-large-v3-peft-swiss-german" # Use the same model ID as before.
47
  peft_config = PeftConfig.from_pretrained(peft_model_id)
48
  model = WhisperForConditionalGeneration.from_pretrained(
49
  peft_config.base_model_name_or_path, load_in_8bit=True, device_map="auto"
 
52
  model.config.use_cache = True
53
 
54
 
 
 
 
 
 
 
 
 
55
  from transformers import AutomaticSpeechRecognitionPipeline
56
  import torch
57
  pipe = AutomaticSpeechRecognitionPipeline(model=model, tokenizer=tokenizer, feature_extractor=feature_extractor)
58
 
59
  with torch.cuda.amp.autocast():
60
+ result = pipe(r"L:\random\audio.mp3", generate_kwargs={"language": "german"})
61
  print(result["text"])
62
+ ```
63
 
64
 
65