In partnership with

Gamechanger: Qwen3 released!

The TLDR
Qwen3 introduces a breakthrough hybrid architecture for language models, allowing it to switch between fast replies and deep thinking modes. The flagship Qwen3-235B-A22B rivals top models like Gemini-2.5-Pro while being more resource-efficient, multilingual, and open-source under Apache 2.0.

The AI race reaches new heights! With the release of Qwen3, a groundbreaking concept is establishing itself in the world of large language models: language models that can switch between quick responses and deep thinking. The flagship Qwen3-235B-A22B successfully competes with industry giants such as DeepSeek-R1, o1, and Gemini-2.5-Pro – and does so with significantly more efficient use of resources.

The innovation lies in the hybrid approach: in “Thinking Mode,” the model works step by step on complex problems, while in “Non-Thinking Mode,” quick answers are provided for simpler questions. This flexibility enables demand-driven control of “thinking power” and optimizes the balance between efficiency and quality.

With support for 119 languages and improved capabilities in the areas of code and agent interaction, Qwen3 is positioning itself as a versatile tool for developers and researchers. Particularly noteworthy: The family includes both MoE (mixture-of-experts) models and dense models of various sizes – all available under the Apache 2.0 license.

Will this hybrid approach enable us to reach the next milestone on the road to AGI? The combination of scaled pre-training and an innovative thinking concept could be just the right step.

Why it matters: Qwen3's hybrid thinking architecture marks a significant advance for practical AI applications, as it is the first to truly combine on-demand thinking with scalable performance. This balance between computational efficiency and depth of thought could pave the way for more responsible, accessible AI systems that are capable of both rapid response and deep analysis.

Learn AI in 5 minutes a day

This is the easiest way for a busy person wanting to learn AI in as little time as possible:

  1. Sign up for The Rundown AI newsletter

  2. They send you 5-minute email updates on the latest AI news and how to use it

  3. You learn how to become 2x more productive by leveraging AI

Graph of the Day

“To develop the hybrid model capable of both step-by-step reasoning and rapid responses, we implemented a four-stage training pipeline. This pipeline includes: (1) long chain-of-thought (CoT) cold start, (2) reasoning-based reinforcement learning (RL), (3) thinking mode fusion, and (4) general RL.” (Qwen 3)

Sparse Attention in LLMs: Balancing Speed, Accuracy, and Scale in Long-Text Processing

Researchers have conducted a comprehensive analysis of “sparse attention” in LLMs – a technique for processing long texts more efficiently. What is new is the systematic investigation of the trade-offs between speed and accuracy across many models and tasks. The key finding: for very long sequences, large, highly “sparse” models outperform smaller, dense models with the same computing budget. However, sparseness is not a universal solution and carries performance risks for specific tasks. It is relevant for processing huge amounts of data, but requires careful, application-specific consideration.

ViSMaP: Meta-Prompting AI Unlocks Scalable, Annotation-Free Video Summarization

Automatically summarize long videos – without expensive data? ViSMaP makes it possible! This new AI system cleverly uses multiple language models (LLMs) via “meta-prompting.” It generates amazingly good, automatic summaries for hours of video from descriptions of short clips – without any human annotation. The performance is comparable to that of elaborately trained methods. This is relevant for analyzing huge video archives and complex content, from social media to film, and circumvents previous data bottlenecks

REFVNLI: A Unified Metric for Evaluating Text-and-Template-Based Image Generation

AI should generate images based on text and a specific template image (e.g., “draw this cat on a tree”). However, evaluating such models is difficult. REFVNLI is a new, cost-efficient AI metric that automatically evaluates both: Was the text description implemented? And: Is the correct subject (e.g., exactly this cat) recognizable? What is new is the combined, reliable evaluation in a single step. This enables advances in personalized image generation and consistent AI creations, which are important for media and design.

Poll of the Day

Will China beat the USA in the AI-race?

Login or Subscribe to participate

In The News

Hugging Face Unveils $100 AI-Powered Robotic Arm as It Expands Into Robotics

Hugging Face has launched the SO-101, a $100 programmable, 3D-printable robotic arm that builds on the success of last year’s SO-100. Developed with The Robot Studio and other partners, the SO-101 is easier to assemble and features upgraded motors and a camera for reinforcement learning-based task training. It can perform simple chores like picking and placing objects, making it ideal for AI hobbyists and developers. The launch comes as Hugging Face expands deeper into robotics, including the acquisition of Pollen Robotics and plans to distribute its humanoid robot, Reachy 2.

Grok 3.5 Beta Launches Next Week With Advanced First-Principles Reasoning

Grok 3.5 enters early beta next week, available exclusively to SuperGrok subscribers. The AI can reason from first principles, enabling it to answer complex technical questions in fields like rocket science and electrochemistry—even when answers aren't found online.

Mark Carney and Liberal Party Win Canadian Election Amid AI-Era Policy Focus

Prime Minister Mark Carney and the Liberal Party have secured victory in the Canadian election. Carney is noted for his candid approach to the future of work in the age of AI, raising hopes for forward-thinking leadership on emerging tech challenges.

Quote of the Day

Hi All,

Thank you for reading. We would be delighted if you shared the newsletter with your friends! We look forward to expanding the newsletter in the future with even more specialized topics. Until then, follow us on social media to stay up to date.

Cheers,
Dan

Reply

or to participate

Keep Reading

No posts found