nickmalhotra commited on
Commit
aee0eac
1 Parent(s): 4144863

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -348,6 +348,11 @@ The pre-training and fine-tuning of Project Indus LLM were conducted on high-per
348
  - **Nodes and GPUs**: Utilization of six nodes, each equipped with eight NVIDIA A100 GPUs. These GPUs are state-of-the-art for machine learning tasks and provide the necessary computational power to handle the large volumes of data and complex model architectures.
349
  - **Memory and Storage**: Each node was equipped with ample memory and storage to handle the datasets and model parameters efficiently. Specific configurations included 40 GB of GPU memory per card, essential for training large models.
350
 
 
 
 
 
 
351
  ##### Software
352
 
353
  The software environment was crucial for efficiently training and running the model. Key components included:
 
348
  - **Nodes and GPUs**: Utilization of six nodes, each equipped with eight NVIDIA A100 GPUs. These GPUs are state-of-the-art for machine learning tasks and provide the necessary computational power to handle the large volumes of data and complex model architectures.
349
  - **Memory and Storage**: Each node was equipped with ample memory and storage to handle the datasets and model parameters efficiently. Specific configurations included 40 GB of GPU memory per card, essential for training large models.
350
 
351
+
352
+ Inference performance was tested on GPU as well as CPU.
353
+ - **GPU**: On GPU NVIDIA GeForce RTX 3070 we have seen for 250-350 tokens inference time around ~5-10s.
354
+ - **CPU**: On Intel CPU Xeon(R) Platinum 8580 we have seen performance comparable to GPU with throughput of > 30 token/second.
355
+
356
  ##### Software
357
 
358
  The software environment was crucial for efficiently training and running the model. Key components included: