FLUX.2
FLUX.2 is built for real production workflows, delivering high-quality visuals while maintaining character, product, and style consistency across multiple reference images. It handles structured prompts, brand-safe layouts, complex text rendering, and detailed logos with precision. The model supports multi-reference inputs, editing at up to 4 megapixels, and generates both photorealistic scenes and highly stylized compositions. With a focus on reliability, FLUX.2 processes real-world creative tasks—such as infographics, product shots, and UI mockups—with exceptional stability. It represents Black Forest Labs’ open-core approach, pairing frontier-level capability with open-weight models that invite experimentation. Across its variants, FLUX.2 provides flexible options for studios, developers, and researchers who need scalable, customizable visual intelligence.
Learn more
Trinity-Large-Thinking
Trinity Large Thinking is a frontier open source reasoning model developed by Arcee AI, designed specifically for complex, multi-step problem solving and autonomous agent workflows that require long-horizon planning and tool use. Built on a sparse Mixture-of-Experts architecture with roughly 400 billion total parameters but only about 13 billion active per token, the model achieves high efficiency while maintaining strong reasoning performance across tasks such as mathematical problem solving, code generation, and multi-step analysis. It introduces extended chain-of-thought reasoning capabilities, allowing the model to generate intermediate “thinking traces” before producing final answers, which improves accuracy and reliability in complex scenarios. Trinity Large Thinking supports a very large context window of up to 262K tokens, enabling it to process long documents, maintain state across extended interactions, and operate effectively in continuous agent loops.
Learn more
EXAONE Deep
EXAONE Deep is a series of reasoning-enhanced language models developed by LG AI Research, featuring parameter sizes of 2.4 billion, 7.8 billion, and 32 billion. These models demonstrate superior capabilities in various reasoning tasks, including math and coding benchmarks. Notably, EXAONE Deep 2.4B outperforms other models of comparable size, EXAONE Deep 7.8B surpasses both open-weight models of similar scale and the proprietary reasoning model OpenAI o1-mini, and EXAONE Deep 32B shows competitive performance against leading open-weight models. The repository provides comprehensive documentation covering performance evaluations, quickstart guides for using EXAONE Deep models with Transformers, explanations of quantized EXAONE Deep weights in AWQ and GGUF formats, and instructions for running EXAONE Deep models locally using frameworks like llama.cpp and Ollama.
Learn more
GLM-5.1
GLM-5.1 is the latest iteration of Z.ai’s GLM series, designed as a frontier-level, agent-oriented AI model optimized for coding, reasoning, and long-horizon workflows. It builds on the GLM-5 architecture, which uses a Mixture-of-Experts (MoE) design to deliver high performance while keeping inference costs efficient, and is part of a broader push toward open-weight, developer-accessible models. A core focus of GLM-5.1 is enabling agentic behavior, meaning it can plan, execute, and iterate across multi-step tasks rather than simply responding to single prompts. It is specifically designed to handle complex workflows such as debugging code, navigating repositories, and executing chained operations with sustained context. Compared to earlier models, GLM-5.1 improves reliability in long interactions, maintaining coherence across extended sessions and reducing breakdowns in multi-step reasoning.
Learn more