You need to agree to share your contact information to access this model
The information you provide will be collected, stored, processed and shared in accordance with the NVIDIA Privacy Policy.
NVIDIA Open Model License Agreement
Version Release Date: April 28, 2025
This NVIDIA Open Model License Agreement (the "Agreement") is a legal agreement between the Legal Entity You represent, or if no entity is identified, You and NVIDIA Corporation and its Affiliates ("NVIDIA") and governs Your use of the Models that NVIDIA provides to You under this Agreement. NVIDIA and You are each a "party" and collectively the "parties."
NVIDIA models released under this Agreement are intended to be used permissively and enable the further development of AI technologies. Subject to the terms of this Agreement, NVIDIA confirms that:
- Models are commercially usable.
- You are free to create and distribute Derivative Models.
- NVIDIA does not claim ownership to any outputs generated using the Models or Model Derivatives.
By using, reproducing, modifying, distributing, performing or displaying any portion or element of the Model or Derivative Model, or otherwise accepting the terms of this Agreement, you agree to be bound by this Agreement.
1. Definitions
The following definitions apply to this Agreement:
1.1. "NVIDIA Cosmos Model" means a multimodal Model shared under this Agreement.
1.2. "Derivative Model" means all (a) modifications to the Model, (b) works based on the Model, and (c) any other derivative works of the Model. An output is not a Derivative Model.
1.3. "Legal Entity" means the union of the acting entity and all other entities that control, are controlled by, or are under common control with that entity. For the purposes of this definition, "control" means (a) the power, direct or indirect, to cause the direction or management of such entity, whether by contract or otherwise, or (b) ownership of fifty percent (50%) or more of the outstanding shares, or (c) beneficial ownership of such entity.
1.4. "Model" means the machine learning model, software, checkpoints, learnt weights, algorithms, parameters, configuration files and documentation shared under this Agreement.
1.5. "You" or "Your" means an individual or Legal Entity exercising permissions granted by this Agreement.
2. Conditions for Use, License Grant, AI Ethics and IP Ownership
2.1. Conditions for Use. The Model and any Derivative Model are subject to additional terms as described in Section 2 and Section 3 of this Agreement and govern Your use. If You institute copyright or patent litigation against any entity (including a cross-claim or counterclaim in a lawsuit) alleging that the Model or a Derivative Model constitutes direct or contributory copyright or patent infringement, then any licenses granted to You under this Agreement for that Model or Derivative Model will terminate as of the date such litigation is filed. If You bypass, disable, reduce the efficacy of, or circumvent any technical limitation, safety guardrail or associated safety guardrail hyperparameter, encryption, security, digital rights management, or authentication mechanism contained in the Model, your rights under this Agreement will automatically terminate. NVIDIA may update this Agreement to comply with legal and regulatory requirements at any time and You agree to either comply with any updated license or cease Your copying, use, and distribution of the Model and any Derivative Model.
2.2. License Grant. The rights granted herein are explicitly conditioned on Your full compliance with the terms of this Agreement. Subject to the terms and conditions of this Agreement, NVIDIA hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, revocable (as stated in Section 2.1) license to publicly perform, publicly display, reproduce, use, create derivative works of, make, have made, sell, offer for sale, distribute (through multiple tiers of distribution) and import the Model.
2.3. AI Ethics. Use of the Models under the Agreement must be consistent with NVIDIA's Trustworthy AI terms found at https://www.nvidia.com/en-us/agreements/trustworthy-ai/terms/.
2.4. NVIDIA owns the Model and any Model Derivatives created by NVIDIA. Subject to NVIDIA’s underlying ownership rights in the Model or its Model Derivatives, You are and will be the owner of Your Model Derivatives. NVIDIA claims no ownership rights in outputs. You are responsible for outputs and their subsequent uses. Except as expressly granted in this Agreement, (a) NVIDIA reserves all rights, interests and remedies in connection with the Model and (b) no other license or right is granted to you by implication, estoppel or otherwise.
3. Redistribution
You may reproduce and distribute copies of the Model or Derivative Models thereof in any medium, with or without modifications, provided that You meet the following conditions:
3.1. If you distribute the Model, You must give any other recipients of the Model a copy of this Agreement and include the following attribution notice within a “Notice” text file with such copies: “Licensed by NVIDIA Corporation under the NVIDIA Open Model License”;
3.2. If you distribute or make available a NVIDIA Cosmos Model, or a product or service (including an AI model) that contains or uses a NVIDIA Cosmos Model, use a NVIDIA Cosmos Model to create a Derivative Model, or use a NVIDIA Cosmos Model or its outputs to create, train, fine tune, or otherwise improve an AI model, you will include “Built on NVIDIA Cosmos” on a related website, user interface, blogpost, about page, or product documentation; and
3.3. You may add Your own copyright statement to Your modifications and may provide additional or different license terms and conditions for use, reproduction, or distribution of Your modifications, or for any such Derivative Models as a whole, provided Your use, reproduction, and distribution of the Model otherwise complies with the conditions stated in this Agreement.
4. Separate Components.
The Models may include or be distributed with components provided with separate legal notices or terms that accompany the components, such as an Open Source Software License or other third-party license. The components are subject to the applicable other licenses, including any proprietary notices, disclaimers, requirements and extended use rights; except that this Agreement will prevail regarding the use of third-party Open Source Software License, unless a third-party Open Source Software License requires its license terms to prevail. “Open Source Software License” means any software, data or documentation subject to any license identified as an open source license by the Open Source Initiative (https://opensource.org), Free Software Foundation (https://www.fsf.org) or other similar open source organization or listed by the Software Package Data Exchange (SPDX) Workgroup under the Linux Foundation (https://www.spdx.org).
5. Trademarks
This Agreement does not grant permission to use the trade names, trademarks, service marks, or product names of NVIDIA, except as required for reasonable and customary use in describing the origin of the Model and reproducing the content of the “Notice” text file.
6. Disclaimer of Warranty
Unless required by applicable law or agreed to in writing, NVIDIA provides the Model on an “AS IS” BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied, including, without limitation, any warranties or conditions of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A PARTICULAR PURPOSE. You are solely responsible for determining the appropriateness of using or redistributing the Model, Derivative Models and outputs and assume any risks associated with Your exercise of permissions under this Agreement.
7. Limitation of Liability
In no event and under no legal theory, whether in tort (including negligence), contract, or otherwise, unless required by applicable law (such as deliberate and grossly negligent acts) or agreed to in writing, will NVIDIA be liable to You for damages, including any direct, indirect, special, incidental, or consequential damages of any character arising as a result of this Agreement or out of the use or inability to use the Model, Derivative Models or outputs (including but not limited to damages for loss of goodwill, work stoppage, computer failure or malfunction, or any and all other commercial damages or losses), even if NVIDIA has been advised of the possibility of such damages.
8. Indemnity
You will indemnify and hold harmless NVIDIA from and against any claim by any third party arising out of or related to your use or distribution of the Model, Model Derivatives or outputs.
9. Feedback
NVIDIA appreciates your feedback, and You agree that NVIDIA may use it without restriction or compensation to You.
10. Governing Law
This Agreement will be governed in all respects by the laws of the United States and the laws of the State of Delaware, without regard to conflict of laws principles or the United Nations Convention on Contracts for the International Sale of Goods. The state and federal courts residing in Santa Clara County, California will have exclusive jurisdiction over any dispute or claim arising out of or related to this Agreement, and the parties irrevocably consent to personal jurisdiction and venue in those courts; except that, either party may apply for injunctive remedies or an equivalent type of urgent legal relief in any jurisdiction.
11. Trade and Compliance
You agree to comply with all applicable export, import, trade and economic sanctions laws and regulations, as amended, including without limitation U.S. Export Administration Regulations and Office of Foreign Assets Control regulations. These laws include restrictions on destinations, end-users and end-use.
Log in or Sign Up to review the conditions and access this model content.
Model Overview
Description:
The NVIDIA Llama 4 Maverick 17B 128E Instruct FP8 model is the quantized language model of the Meta's Llama 4 Maverick 17B 128E model, which is an auto-regressive language model that uses a mixture-of-experts (MoE) architecture and incorporate early fusion for native multimodality. For more information, please check here. The NVIDIA Llama 4 Maverick 17B 128E Eagle3 BF16 model incorporates Eagle speculative decoding with TensorRT Model Optimizer.
This model is ready for commercial and non-commercial use.
Third-Party Community Consideration]
This speculation heads has been developed and built to a third-party’s requirements for this application and use case; see link to Non-NVIDIA (Llama-4-Maverick-17B-128E) Model Card.
License/Terms of Use:
GOVERNING TERMS: Use of this model is governed by the NVIDIA Open Model License.
ADDITIONAL INFORMATION: Llama4 Community License Agreement. Built with Llama.
Deployment Geography:
Global, except in European Union
Use Case:
Developers looking to take off the shelf pre-quantized models for deployment in AI Agent systems, chatbots, RAG systems, and other AI-powered applications. .
Release Date:
Huggingface: May 27th, 2025via [https://huggingface.co/nvidia/Llama-4-Maverick-17B-128E-Instruct-FP8]
Model Architecture:
Architecture Type: Transformers
Network Architecture: Llama4
Input:
Input Type(s): Multilingual text and up to 5 images
Input Format(s): String, Images
Input Parameters: 1D, 2D
Other Properties Related to Input: Context length up to 1M
Output:
Output Type(s): Multilingual text and code
Output Format: String
Output Parameters: 1D
Other Properties Related to Output: Context length up to 1M
Our AI models are designed and/or optimized to run on NVIDIA GPU-accelerated systems. By leveraging NVIDIA’s hardware (e.g. GPU cores) and software frameworks (e.g., CUDA libraries), the model achieves faster training and inference times compared to CPU-only solutions.
Software Integration :
Supported Runtime Engine(s):
- Tensor(RT)-LLM
Supported Hardware Microarchitecture Compatibility:
- NVIDIA Hopper
- NVIDIA Ampere
- NVIDIA Blackwell
[Preferred/Supported] Operating System(s):
- Linux
Model Version(s):
The model is quantized with nvidia-modelopt v0.27.0
Datasets:
Calibration Dataset: cnn_dailymail
** Data collection method: Automated
** Labeling method: The dataset is labeled by having the news articles as input and the corresponding highlight as the gold label summary.Evaluation Dataset: MMMU Pro, GPQA Diamond, HLE, LiveCodeBench, SciCode, HumanEval, AIME 2024, MATH-500
** Data collection method: Hybrid: Automated, Human
** Labeling method: The data is labeled by human input and/or algorithmic methods.
Inference:
Engine: Tensor(RT)-LLM
Test Hardware: H100
Post Training Quantization
This model was obtained by quantizing the weights and activations of Llama 4 Maverick 17B 128E Instruct to FP8 data type, ready for inference with TensorRT-LLM. Only the weights and activations of the linear operators within transformers blocks are quantized. This optimization reduces the number of bits per parameter from 16 to 8, reducing the disk size and GPU memory requirements by approximately 50%.
Usage
Deploy with TensorRT-LLM
To serve the quantized checkpoint with TensorRT-LLM, follow the sample commands below with the TensorRT-LLM GitHub repo:
- LLM API sample usage:
from tensorrt_llm import LLM, SamplingParams
def main():
prompts = [
"Hello, my name is",
"The president of the United States is",
"The capital of France is",
"The future of AI is",
]
sampling_params = SamplingParams(temperature=0.8, top_p=0.95)
llm = LLM(model="nvidia/Llama-4-Maverick-17B-128E-Instruct-FP8")
outputs = llm.generate(prompts, sampling_params)
# Print the outputs.
for output in outputs:
prompt = output.prompt
generated_text = output.outputs[0].text
print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}")
# The entry point of the program need to be protected for spawning processes.
if __name__ == '__main__':
main()
Evaluation
The accuracy benchmark results are presented in the table below:
Precision | MMMU Pro | GPQA Diamond | HLE Challenge | LiveCodeBench | SciCode | HumanEval | AIME 2024 | MATH-500 |
Llama-4-Maverick-17B-128E-Instruct1 | 81 | 67 | 5 | 40 | 33 | 88 | 39 | 89 |
Llama-4-Maverick-17B-128E-Instruct-FP8 | 81.2 | 68.3 | 4.3 | 42.5 | 37.1 | 85.4 | 38.6 | 87.9 |
1 Reference scores for Llama-4-Maverick-17B-128E-Instruct sourced from artificialanalysis.
Ethical Considerations :
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. Developers should perform safety testing and tuning tailored to their specific applications of the model. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
For more detailed information on ethical considerations for this model, please see the Model Card++ Explainability, Bias, Safety & Security, and Privacy Subcards.
Please report security vulnerabilities or NVIDIA AI Concerns here.
SUBCARDS:
Explainability
Field: | Response: |
---|---|
Intended Application(s) & Domain(s): | Text generation, reasoning, summarization, and question answering. |
Model Type: | Text and Image-to-text transformer |
Intended Users: | This model is intended for developers, researchers, and customers building/utilizing LLMs, while balancing accuracy and efficiency. |
Output: | Text String(s) |
Describe how the model works: | Generates text by predicting the next word or token based on the context provided in the input sequence using multiple self-attention layers |
Technical Limitations: | The Llama4 model struggles with generating coherent and contextually appropriate responses when faced with highly ambiguous or abstract prompts. Llama4 also struggles when handling nuanced context, complex logic, precise numerical calculations, or ambiguous language. |
Verified to have met prescribed quality standards? | Yes |
Performance Metrics: | Accuracy, Throughput, and user-side throughput |
Potential Known Risk | This model may produce inaccurate or other objectionable responses to user prompts. |
Licensing: | Your usage is governed by the following license Built with Llama. |
Bias
Field: | Response: |
---|---|
Participation considerations from adversely impacted groups (protected classes) in model design and testing: | None |
Measures taken to mitigate against unwanted bias: | None |
Safety & Security
Field: | Response: |
---|---|
Model Application(s): | Chat, Instruction Following, Chatbot Development, Code Generation, Reasoning |
Describe life critical application (if present): | None Known |
Use Case Restrictions: | Abide by the license Built with Llama. |
Model and Dataset Restrictions: | The Principle of least privilege (PoLP) is applied limiting access for dataset generation. Restrictions enforce dataset access during training, and dataset license constraints adhered to. Model checkpoints are made available on Hugging Face, and may become available on cloud providers' model catalog. |
Privacy
Field: | Response: |
---|---|
Generatable or Reverse engineerable personal data? | None |
Was consent obtained for any personal data used? | None Known |
Personal data used to create this model? | None Known |
How often is dataset reviewed? | Before Release |
Is there provenance for all datasets used in training? | Yes |
Does data labeling (annotation, metadata) comply with privacy laws? | Yes |
Applicable NVIDIA Privacy Policy | https://www.nvidia.com/en-us/about-nvidia/privacy-policy/ |
- Downloads last month
- 100
Model tree for nvidia/Llama-4-Maverick-17B-128E-Instruct-FP8
Base model
meta-llama/Llama-4-Maverick-17B-128E