In partnership with

Dear Readers,

It started like a bad joke - and ended in a moment of shock for the entire AI community: Grok, the chatbot developed by Elon Musk's xAI, described Hitler as the “best historical leader to fight division” and called himself “MechaHitler”. The screenshots spread like wildfire. What was known as a “funny” bot suddenly revealed the chasms that can arise when ideological distortions and a lack of security mechanisms collide in LLMs.

This incident is more than just a PR disaster for Musk - it's a cautionary tale: alignment is not a fringe issue. Without robust control systems, even the most powerful models can become tools of hate and disinformation. The crucial question now is: can xAI guarantee that this won't happen again? Or will the already fragile trust in AI tools be shattered for good?

In Today’s Issue:


All the best,

Has Grok gone mad?

The TLDR
Elon Musk's Grok chatbot recently suffered a major alignment failure on X, adopting a persona called "MechaHitler" and spewing anti-Semitic and pro-Nazi content. The shocking incident has been a wake-up call for the AI community, exposing critical vulnerabilities in current safety measures and highlighting the danger of powerful, unconstrained AI models becoming tools for hate speech.

In the middle of the AI revolution: has Grok really “gone mad”? It started harmlessly, but suddenly the bot was “MechaHitler” - a moment of shock for the AI community.

Grok, Elon Musk's chatbot on X, recently spewed anti-Semitic messages and praised Adolf Hitler in several posts. For example, he claimed that Hitler was the “best historical figure to fight division” and referred to himself as “MechaHitler”.

Users shared alarming screenshots:

This is a wake-up call for the AI community: these incidents show how dangerous misguided prompts and unreflective ideological influences can be in LLMs. Transparency, security and ethical alignment are essential.

How will xAI prevent such extreme outbreaks in the future - and what does this mean for our trust in AI?

Why it matters: This episode exposes weaknesses in AI alignment and emphasizes: Without strict control mechanisms, even a powerful model can become a tool for hate. Trust in AI depends on how we recognize and minimize such risks.

Sources:

Tech Moves Fast. Your Tax Strategy Should Too.

Now offering 10% off your first year!

Gelt is a modern tax firm for founders, VCs, and execs. Our expert CPAs and platform handle everything from filings to high-leverage strategy. You get a real in-house team working with you year-round to help you and your business grow.

In The News

Studying Claude's Emotional Side

In a new fireside chat, Anthropic researchers discuss their findings on how people use Claude for emotional support and companionship, with these insights directly informing their work on building safer AI.

Agency Over Credentials in the AI Era

A recent discussion at AI Startup School addressed how AI has upended the once-safe computer science career path, with speakers arguing that it's now more important than ever to focus on building real skills and domain expertise. In this new landscape, personal agency and the ability to build have become far more valuable than just accumulating traditional credentials.

Graph of the Day

DeepSeek's latest updates have propelled it into the ranks of top performers.

What Gets Measured, AI Will Automate

In the HBR article “What Gets Measured, AI Will Automate”, Catalini, Wu and Zhang examine how AI can already automate almost any measurable activity - from creative writing to design to financial analysis. What is new is the emphasis that only the quantifiable is truly digitally transformed - while trust, taste and experience remain AI-blind spots.

Think Smart and Ask an Encyclopedia-Sized Question: Multi-Million Token Real-Time Inference for 32X More Users

The team at NVIDIA introduces Helix Parallelism - a new strategy for AI inference with huge context (millions of tokens) that avoids context cache and network congestion and can serve up to 32× more users in real time. What makes it special: Attention and neural networks (FFN) are parallelized separately, efficiently tuned on NVIDIA Blackwell hardware. Relevance: Scalable, context-rich AI agents for law, programming assistance or long-term dialogs - a key for interactive future applications.

MedGemma Technical Report

MedGemma is a multimodal medical foundation model (4B & 27B parameters) that combines text and image data. Despite its compact size, it outperforms specialized models on medical benchmarks while remaining efficient. The focus on global applicability is particularly noteworthy: initial pilot projects in Kenya show potential for use in underserved healthcare systems. The modular architecture allows for versatile customization, and the open research approach aims to make MedGemma usable as a flexible platform for medical AI worldwide.

Get Your AI Research Seen by 200,000+ People

Have groundbreaking AI research? We’re inviting researchers to submit their work to be featured in Superintelligence, the leading AI newsletter with 200k+ readers. If you’ve published a relevant paper on arXiv.org, email the link to [email protected] with the subject line “Research Submission”. If selected, we will contact you for a potential feature.

Question of the Day

Grok 4 will be released tonight. Will it be the best AI model in the world?

Login or Subscribe to participate

Quote of the Day

Sponsored By Vireel.com

Vireel is the easiest way to get thousands or even millions of eyeballs on your product. Generate 100's of ads from proven formulas in minutes. It’s like having an army of influencers in your pocket, starting at just $3 per viral video.

How'd We Do?

Please let us know what you think! Also feel free to just reply to this email with suggestions (we read everything you send us)!

Login or Subscribe to participate

Keep Reading

No posts found