Tencent-Hunyuan-Large is the flagship open-source large language model family from Tencent Hunyuan, offering both pre-trained and instruct (fine-tuned) variants. It is designed with long-context capabilities, quantization support, and high performance on benchmarks across general reasoning, mathematics, language understanding, and Chinese / multilingual tasks. It aims to provide competitive capability with efficient deployment and inference. FP8 quantization support to reduce memory usage (~50%) while maintaining precision. High benchmarking performance on tasks like MMLU, MATH, CMMLU, C-Eval, etc.
Features
- Long context window support: up to 256K tokens in pretrain; 128K tokens for instruct models
- FP8 quantization support to reduce memory usage (~50%) while maintaining precision
- Expert-specific learning rate scaling in training (for mixture or expert architectures)
- High benchmarking performance on tasks like MMLU, MATH, CMMLU, C-Eval, etc.
- Hugging Face format compatibility for fine-tuning / inference using frameworks like hf-deepspeed, plus support for flash attention, efficient operators (TRT-LLM)
- Throughput and efficiency improvements: TRT-LLM backend surpasses vLLM by ~30 %, quantized/inference optimizations included
Follow Tencent-Hunyuan-Large
Other Useful Business Software
MongoDB Atlas runs apps anywhere
MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
Rate This Project
Login To Rate This Project
User Reviews
Be the first to post a review of Tencent-Hunyuan-Large!