NeMo
okuchaiev Chris-Alexiuk commited on
Commit
fd0705b
·
verified ·
1 Parent(s): 08f18d9

Update README.md (#3)

Browse files

- Update README.md (71a1b51c83e15e2f2e831a0b18f7ad34490bfbfc)


Co-authored-by: Chris Alexiuk <[email protected]>

Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -24,9 +24,9 @@ Throughout the alignment process, we relied on only approximately 20K human-anno
24
  This results in a model that is aligned for human chat preferences, improvements in mathematical reasoning, coding and instruction-following, and is capable of generating high quality synthetic data for a variety of use cases.
25
 
26
  Under the NVIDIA Open Model License, NVIDIA confirms:
27
- Models are commercially usable.
28
- You are free to create and distribute Derivative Models.
29
- NVIDIA does not claim ownership to any outputs generated using the Models or Derivative Models.
30
 
31
  ### License:
32
 
@@ -310,9 +310,9 @@ Evaluated using the CantTalkAboutThis Dataset as introduced in the [CantTalkAbou
310
  ### Adversarial Testing and Red Teaming Efforts
311
 
312
  The Nemotron-4 340B-Instruct model underwent extensive safety evaluation including adversarial testing via three distinct methods:
313
- [Garak](https://docs.garak.ai/garak), is an automated LLM vulnerability scanner that probes for common weaknesses, including prompt injection and data leakage.
314
- [AEGIS](https://arxiv.org/pdf/2404.05993), is a content safety evaluation dataset and LLM based content safety classifier model, that adheres to a broad taxonomy of 13 categories of critical risks in human-LLM interactions.
315
- Human Content Red Teaming leveraging human interaction and evaluation of the models' responses.
316
 
317
  ### Limitations
318
 
 
24
  This results in a model that is aligned for human chat preferences, improvements in mathematical reasoning, coding and instruction-following, and is capable of generating high quality synthetic data for a variety of use cases.
25
 
26
  Under the NVIDIA Open Model License, NVIDIA confirms:
27
+ - Models are commercially usable.
28
+ - You are free to create and distribute Derivative Models.
29
+ - NVIDIA does not claim ownership to any outputs generated using the Models or Derivative Models.
30
 
31
  ### License:
32
 
 
310
  ### Adversarial Testing and Red Teaming Efforts
311
 
312
  The Nemotron-4 340B-Instruct model underwent extensive safety evaluation including adversarial testing via three distinct methods:
313
+ - [Garak](https://docs.garak.ai/garak), is an automated LLM vulnerability scanner that probes for common weaknesses, including prompt injection and data leakage.
314
+ - [AEGIS](https://arxiv.org/pdf/2404.05993), is a content safety evaluation dataset and LLM based content safety classifier model, that adheres to a broad taxonomy of 13 categories of critical risks in human-LLM interactions.
315
+ - Human Content Red Teaming leveraging human interaction and evaluation of the models' responses.
316
 
317
  ### Limitations
318