Model Card for Model ID

This modelcard aims to be a base template for new models. It has been generated using this raw template.

Model Details

Model Description

DeepSeek-V2-Lite Quantized to NVFP4 Using TensorRT-Model-Optimizer. The model has 15.6B parameters in total, in Nvidia float precision 4 bit. Require Blackwell GPU, NVFP4 supported inference engine(if use with inference engine)

  • Developed by: Krisakorn Chanthasang
  • Model type: LLM TextGeneration
  • Language(s) (NLP): Chinese English Thai
  • License: Apache 2.0
  • Quantized from model: DeepSeek-V2-Lite

Model Information

Read https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite for Basic model information

How to Get Started with the Model

GPU requirement: Blackwell Series

Hardware

Quantization Hardware: Nvidia RTX PRO 6000 Blackwell Workstation

Downloads last month
23
Safetensors
Model size
8B params
Tensor type
BF16
·
F8_E4M3
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Beambutbetter/Deepseek-V2-Lite-16B-NVFP4

Quantized
(18)
this model