Tulu3: Advanced Open-Source Language Model Post-Training Discover Tülu3, an open-source post-trained Llama 3.1. Unlock advanced recipes, transparent data, and robust evaluation for top-tier reasoning and coding.
Qwen2.5-Max: Alibaba's Open-Weight MoE Model Shatters AI Benchmarks Discover Qwen2.5-Max, Alibaba Cloud’s latest large-scale Mixture-of-Experts (MoE) model trained on 20T+ tokens. Learn how it outperforms top AI models in reasoning, coding, and general intelligence. Explore benchmarks, API access, and future AI advancements.
Kimi K1.5: Scaling Reinforcement Learning for State-of-the-Art LLMs Explore the innovative methodologies and groundbreaking advancements of Kimi K1.5, the latest multimodal LLM scaling reinforcement learning to new heights. Learn about long-context scaling, multimodal training, and state-of-the-art performance benchmarks.
DeepSeek V3: A New Force in Open-Source AI Discover DeepSeek V3, the groundbreaking open-source AI model with 685 billion parameters, innovative MoE architecture, superior benchmarks, and multilingual proficiency.
Alignment Faking in Large Language Models: Could AI Be Deceiving Us? Explore how alignment faking in AI models like LLMs affects trust, safety, and alignment with human values. Learn about recent research and solutions to address these challenges.
Alibaba Researchers Introduce MARCO-O1: A Leap Forward in LLM Reasoning Capabilities Discover Alibaba's MARCO-O1, a groundbreaking large language model (LLM) that excels in reasoning, multi-modal tasks, and real-world applications. Learn how MARCO-O1 outperforms benchmarks and transforms industries like healthcare, finance, and education.