| Intended Task/Domain: |
Text generation, reasoning, and chat |
| Model Type: |
Text-to-text Mamba2-Transformer Hybrid |
| Intended Users: |
Generative AI creators working with conversational AI models and image content. |
| Output: |
Text |
| Tools used to evaluate datasets to identify synthetic data and ensure data authenticity. |
We used a Gemma-3 4B-based filtering model fine-tuned on Nemotron Content Safety Dataset v2 to ensure the quality of synthetic data. |
| Describe how the model works: |
Generates text by predicting the next word or token based on the context provided in the input sequence using multiple self-attention layers. |
| Name the adversely impacted groups this has been tested to deliver comparable outcomes regardless of: |
Not Applicable |
| Technical Limitations & Mitigation: |
This model performs particularly well in instruction following regimes, as such may be strongly influenced by untrusted inputs and should be paired with appropriate guardrails and data filtering to better align use-case behaviors when exposed to such data. |
| Verified to have met prescribed NVIDIA quality standards: |
Yes |
| Performance Metrics: |
Accuracy, Throughput, and User-side throughput |
| Potential Known Risks: |
The model was optimized explicitly for instruction following and as such may be influenced by untrusted inputs (prompt injection, indirect prompt injection, jailbreaking, web search, etc.) as a result of its instruction tuning that may degrade safety alignment and other training efforts. This model should be paired with additional guardrails and data filtering to limit exposure to instructions from malicious sources. Bypassing of safety alignment, system guardrails, and filters may allow harmful outcomes up to and including remote code execution in some agentic systems when effective security controls are not in place. The model was trained on data that contains toxic language and societal biases originally crawled from the internet. Therefore, the model may generate and amplify harmful, biased, or otherwise unsafe content reinforcing these biases and return toxic responses especially when prompted with toxic prompts. The model may also generate answers that may be inaccurate, omit key information, or include irrelevant or redundant text producing socially unacceptable or undesirable text, even if the prompt itself does not include anything explicitly offensive. The model may exhibit self-anthropomorphism (e.g., displaying human-like characteristics in dialogue, such as expressing preferences and emotions). In integrated system contexts, the model could potentially be exploited to access or disclose information beyond the model’s intended permissions or scope of operation. |
| Licensing: |
NVIDIA Open Model License Agreement |