NeMo
jiaqiz commited on
Commit
91adaf3
1 Parent(s): 3872adc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -4
README.md CHANGED
@@ -1,7 +1,8 @@
1
  ---
2
  license: other
3
  license_name: nvidia-open-model-license
4
- license_link: LICENSE
 
5
  ---
6
  ## Nemotron-4-340B-Instruct
7
 
@@ -120,7 +121,7 @@ if response.endswith("<extra_id_1>"):
120
  print(response)
121
  ```
122
 
123
- 2. Given this python script, we will create a bash script, which spins up the inference server within the NeMo container(docker pull nvcr.io/nvidia/nemo:24.01.framework) and calls the python script ``call_server.py``. The bash script ``nemo_inference.sh`` is as follows,
124
 
125
  ```bash
126
  NEMO_FILE=$1
@@ -170,7 +171,7 @@ depends_on () {
170
  ```
171
 
172
 
173
- 3, We can launch the ``nemo_inferece.sh`` with a slurm script defined like below, which starts a 4-node job for the model inference.
174
 
175
  ```bash
176
  #!/bin/bash
@@ -286,4 +287,4 @@ The Nemotron-4 340B-Instruct model underwent extensive safety evaluation includi
286
 
287
  ### Ethical Considerations
288
 
289
- NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. For more detailed information on ethical considerations for this model, please see the Model Card++ Explainability, Bias, Safety & Security, and Privacy Subcards [Insert Link to Model Card++ here]. Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).
 
1
  ---
2
  license: other
3
  license_name: nvidia-open-model-license
4
+ license_link: >-
5
+ https://developer.download.nvidia.com/licenses/nvidia-open-model-license-agreement-june-2024.pdf
6
  ---
7
  ## Nemotron-4-340B-Instruct
8
 
 
121
  print(response)
122
  ```
123
 
124
+ 2. Given this python script, we will create a bash script, which spins up the inference server within the NeMo container (```docker pull nvcr.io/nvidia/nemo:24.01.framework```) and calls the python script ``call_server.py``. The bash script ``nemo_inference.sh`` is as follows,
125
 
126
  ```bash
127
  NEMO_FILE=$1
 
171
  ```
172
 
173
 
174
+ 3, We can launch the ``nemo_inferece.sh`` with a slurm script defined like below, which starts a 2-node job for the model inference.
175
 
176
  ```bash
177
  #!/bin/bash
 
287
 
288
  ### Ethical Considerations
289
 
290
+ NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. For more detailed information on ethical considerations for this model, please see the Model Card++ Explainability, Bias, Safety & Security, and Privacy Subcards [Insert Link to Model Card++ here]. Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).