sander-wood commited on
Commit
d971d28
1 Parent(s): 5b4ca75

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -10,7 +10,7 @@ You can check out the [demo page](https://byte-gpt.github.io/), which includes e
10
  ## Model Description
11
  Traditional deep learning often overlooks bytes, the basic units of the digital world, where all forms of information and operations are encoded and manipulated in binary format. Inspired by the success of next token prediction in natural language processing, we introduce bGPT, a model with next byte prediction to simulate the digital world. bGPT matches specialized models in performance across various modalities, including text, audio, and images, and offers new possibilities for predicting, simulating, and diagnosing algorithm or hardware behaviour. It has almost flawlessly replicated the process of converting symbolic music data, achieving a low error rate of 0.0011 bits per byte in converting ABC notation to MIDI format. In addition, bGPT demonstrates exceptional capabilities in simulating CPU behaviour, with an accuracy exceeding 99.99% in executing various operations. Leveraging next byte prediction, models like bGPT can directly learn from vast binary data, effectively simulating the intricate patterns of the digital world.
12
 
13
- We provide five weights of bGPT on [Hugging Face](https://huggingface.co/sander-wood/bgpt) (check them on "Files and versions") corresponding to each dataset used for pre-training:
14
 
15
  1. **_weights-conversion.pth_**: bGPT pre-trained on IrishMAN for data conversion.
16
  2. **_weights-cpu.pth_**: bGPT pre-trained on CPU states for CPU state modelling.
@@ -44,7 +44,7 @@ To set up the bGPT environment and install the necessary dependencies, follow th
44
  ```
45
  4. Download Pre-trained bGPT Weights (Optional)
46
 
47
- For those interested in starting with pre-trained models, bGPT weights are available on [Hugging Face](https://huggingface.co/sander-wood/bgpt). This step is optional but recommended for users looking to leverage the model's capabilities without training from scratch.
48
 
49
  ## Usage
50
 
 
10
  ## Model Description
11
  Traditional deep learning often overlooks bytes, the basic units of the digital world, where all forms of information and operations are encoded and manipulated in binary format. Inspired by the success of next token prediction in natural language processing, we introduce bGPT, a model with next byte prediction to simulate the digital world. bGPT matches specialized models in performance across various modalities, including text, audio, and images, and offers new possibilities for predicting, simulating, and diagnosing algorithm or hardware behaviour. It has almost flawlessly replicated the process of converting symbolic music data, achieving a low error rate of 0.0011 bits per byte in converting ABC notation to MIDI format. In addition, bGPT demonstrates exceptional capabilities in simulating CPU behaviour, with an accuracy exceeding 99.99% in executing various operations. Leveraging next byte prediction, models like bGPT can directly learn from vast binary data, effectively simulating the intricate patterns of the digital world.
12
 
13
+ We provide five weights of bGPT on [Hugging Face](https://huggingface.co/sander-wood/bgpt/tree/main) corresponding to each dataset used for pre-training:
14
 
15
  1. **_weights-conversion.pth_**: bGPT pre-trained on IrishMAN for data conversion.
16
  2. **_weights-cpu.pth_**: bGPT pre-trained on CPU states for CPU state modelling.
 
44
  ```
45
  4. Download Pre-trained bGPT Weights (Optional)
46
 
47
+ For those interested in starting with pre-trained models, bGPT weights are available on [Hugging Face](https://huggingface.co/sander-wood/bgpt/tree/main). This step is optional but recommended for users looking to leverage the model's capabilities without training from scratch.
48
 
49
  ## Usage
50