Update README.md
Browse files
README.md
CHANGED
@@ -1,7 +1,8 @@
|
|
1 |
---
|
2 |
license: other
|
3 |
license_name: nvidia-open-model-license
|
4 |
-
license_link:
|
|
|
5 |
---
|
6 |
## Nemotron-4-340B-Instruct
|
7 |
|
@@ -120,7 +121,7 @@ if response.endswith("<extra_id_1>"):
|
|
120 |
print(response)
|
121 |
```
|
122 |
|
123 |
-
2. Given this python script, we will create a bash script, which spins up the inference server within the NeMo container(docker pull nvcr.io/nvidia/nemo:24.01.framework) and calls the python script ``call_server.py``. The bash script ``nemo_inference.sh`` is as follows,
|
124 |
|
125 |
```bash
|
126 |
NEMO_FILE=$1
|
@@ -170,7 +171,7 @@ depends_on () {
|
|
170 |
```
|
171 |
|
172 |
|
173 |
-
3, We can launch the ``nemo_inferece.sh`` with a slurm script defined like below, which starts a
|
174 |
|
175 |
```bash
|
176 |
#!/bin/bash
|
@@ -286,4 +287,4 @@ The Nemotron-4 340B-Instruct model underwent extensive safety evaluation includi
|
|
286 |
|
287 |
### Ethical Considerations
|
288 |
|
289 |
-
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. For more detailed information on ethical considerations for this model, please see the Model Card++ Explainability, Bias, Safety & Security, and Privacy Subcards [Insert Link to Model Card++ here]. Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).
|
|
|
1 |
---
|
2 |
license: other
|
3 |
license_name: nvidia-open-model-license
|
4 |
+
license_link: >-
|
5 |
+
https://developer.download.nvidia.com/licenses/nvidia-open-model-license-agreement-june-2024.pdf
|
6 |
---
|
7 |
## Nemotron-4-340B-Instruct
|
8 |
|
|
|
121 |
print(response)
|
122 |
```
|
123 |
|
124 |
+
2. Given this python script, we will create a bash script, which spins up the inference server within the NeMo container (```docker pull nvcr.io/nvidia/nemo:24.01.framework```) and calls the python script ``call_server.py``. The bash script ``nemo_inference.sh`` is as follows,
|
125 |
|
126 |
```bash
|
127 |
NEMO_FILE=$1
|
|
|
171 |
```
|
172 |
|
173 |
|
174 |
+
3, We can launch the ``nemo_inferece.sh`` with a slurm script defined like below, which starts a 2-node job for the model inference.
|
175 |
|
176 |
```bash
|
177 |
#!/bin/bash
|
|
|
287 |
|
288 |
### Ethical Considerations
|
289 |
|
290 |
+
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. For more detailed information on ethical considerations for this model, please see the Model Card++ Explainability, Bias, Safety & Security, and Privacy Subcards [Insert Link to Model Card++ here]. Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).
|