reach-vb HF staff commited on
Commit
bf047d5
β€’
1 Parent(s): 4a6083c

Update README.md (#4)

Browse files

- Update README.md (ae923b050f2e9372f51478d622f9ccf9157659eb)

Files changed (1) hide show
  1. README.md +7 -6
README.md CHANGED
@@ -63,22 +63,23 @@ Try out MusicGen yourself!
63
 
64
  ## πŸ€— Transformers Usage
65
 
66
- You can run MusicGen locally with the πŸ€— Transformers library from version 4.31.0 onwards.
67
 
68
  1. First install the πŸ€— [Transformers library](https://github.com/huggingface/transformers) and scipy:
69
 
70
  ```
71
  pip install --upgrade pip
72
- pip install --upgrade transformers scipy
73
  ```
74
 
75
  2. Run inference via the `Text-to-Audio` (TTA) pipeline. You can infer the MusicGen model via the TTA pipeline in just a few lines of code!
76
 
77
  ```python
78
- from transformers import pipeline
79
  import scipy
 
 
80
 
81
- synthesiser = pipeline("text-to-audio", "facebook/musicgen-stereo-medium")
82
 
83
  music = synthesiser("lo-fi music with a soothing melody", forward_params={"do_sample": True})
84
 
@@ -91,13 +92,13 @@ scipy.io.wavfile.write("musicgen_out.wav", rate=music["sampling_rate"], music=au
91
  from transformers import AutoProcessor, MusicgenForConditionalGeneration
92
 
93
  processor = AutoProcessor.from_pretrained("facebook/musicgen-stereo-medium")
94
- model = MusicgenForConditionalGeneration.from_pretrained("facebook/musicgen-stereo-medium")
95
 
96
  inputs = processor(
97
  text=["80s pop track with bassy drums and synth", "90s rock song with loud guitars and heavy drums"],
98
  padding=True,
99
  return_tensors="pt",
100
- )
101
 
102
  audio_values = model.generate(**inputs, max_new_tokens=256)
103
  ```
 
63
 
64
  ## πŸ€— Transformers Usage
65
 
66
+ You can run MusicGen Stereo models locally with the πŸ€— Transformers library from `main` onward.
67
 
68
  1. First install the πŸ€— [Transformers library](https://github.com/huggingface/transformers) and scipy:
69
 
70
  ```
71
  pip install --upgrade pip
72
+ pip install --upgrade git+https://github.com/huggingface/transformers.git scipy
73
  ```
74
 
75
  2. Run inference via the `Text-to-Audio` (TTA) pipeline. You can infer the MusicGen model via the TTA pipeline in just a few lines of code!
76
 
77
  ```python
 
78
  import scipy
79
+ import torch
80
+ from transformers import pipeline
81
 
82
+ synthesiser = pipeline("text-to-audio", "facebook/musicgen-stereo-medium", torch_dtype=torch.float16, device="cuda")
83
 
84
  music = synthesiser("lo-fi music with a soothing melody", forward_params={"do_sample": True})
85
 
 
92
  from transformers import AutoProcessor, MusicgenForConditionalGeneration
93
 
94
  processor = AutoProcessor.from_pretrained("facebook/musicgen-stereo-medium")
95
+ model = MusicgenForConditionalGeneration.from_pretrained("facebook/musicgen-stereo-medium").to("cuda")
96
 
97
  inputs = processor(
98
  text=["80s pop track with bassy drums and synth", "90s rock song with loud guitars and heavy drums"],
99
  padding=True,
100
  return_tensors="pt",
101
+ ).to("cuda")
102
 
103
  audio_values = model.generate(**inputs, max_new_tokens=256)
104
  ```