mimbres commited on
Commit
ba9248c
·
verified ·
1 Parent(s): aade64b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +13 -7
app.py CHANGED
@@ -81,6 +81,7 @@ def prepare_media(source_path_or_url: os.PathLike,
81
 
82
  for line in iter(process.stdout.readline, ''):
83
  # Filter out unnecessary messages
 
84
  if "www.google.com/device" in line:
85
  hl_text = line.replace("https://www.google.com/device", "\033[93mhttps://www.google.com/device\x1b[0m").split()
86
  hl_text[-1] = "\x1b[31;1m" + hl_text[-1] + "\x1b[0m"
@@ -173,14 +174,19 @@ with gr.Blocks(theme=theme, css=css) as demo:
173
  gr.Markdown(
174
  f"""
175
  ## 🎶YourMT3+: Multi-instrument Music Transcription with Enhanced Transformer Architectures and Cross-dataset Stem Augmentation
176
- ## Model card:
177
  - Model name: `{model_name}`
178
- - Encoder backbone: Perceiver-TF + Mixture of Experts (2/8)
179
- - Decoder backbone: Multi-channel T5-small
180
- - Tokenizer: MT3 tokens with Singing extension
181
- - Dataset: YourMT3 dataset
182
- - Augmentation strategy: Intra-/Cross dataset stem augment, No Pitch-shifting
183
- - FP Precision: BF16-mixed for training, FP16 for inference
 
 
 
 
 
 
184
 
185
  ## Caution:
186
  - For acadmic reproduction purpose, we strongly recommend to use [Colab Demo](https://colab.research.google.com/drive/1AgOVEBfZknDkjmSRA7leoa81a2vrnhBG?usp=sharing) with multiple checkpoints.
 
81
 
82
  for line in iter(process.stdout.readline, ''):
83
  # Filter out unnecessary messages
84
+ print(line)
85
  if "www.google.com/device" in line:
86
  hl_text = line.replace("https://www.google.com/device", "\033[93mhttps://www.google.com/device\x1b[0m").split()
87
  hl_text[-1] = "\x1b[31;1m" + hl_text[-1] + "\x1b[0m"
 
174
  gr.Markdown(
175
  f"""
176
  ## 🎶YourMT3+: Multi-instrument Music Transcription with Enhanced Transformer Architectures and Cross-dataset Stem Augmentation
 
177
  - Model name: `{model_name}`
178
+ <▶model details◀>
179
+ <summary>(Details)</summary>
180
+
181
+ | **Component** | **Details** |
182
+ |--------------------------|--------------------------------------------------|
183
+ | Encoder backbone | Perceiver-TF + Mixture of Experts (2/8) |
184
+ | Decoder backbone | Multi-channel T5-small |
185
+ | Tokenizer | MT3 tokens with Singing extension |
186
+ | Dataset | YourMT3 dataset |
187
+ | Augmentation strategy | Intra-/Cross dataset stem augment, No Pitch-shifting |
188
+ | FP Precision | BF16-mixed for training, FP16 for inference |
189
+ </details>
190
 
191
  ## Caution:
192
  - For acadmic reproduction purpose, we strongly recommend to use [Colab Demo](https://colab.research.google.com/drive/1AgOVEBfZknDkjmSRA7leoa81a2vrnhBG?usp=sharing) with multiple checkpoints.