In partnership with

Dear Readers,
It started like a bad joke - and ended in a moment of shock for the entire AI community: Grok, the chatbot developed by Elon Musk's xAI, described Hitler as the “best historical leader to fight division” and called himself “MechaHitler”. The screenshots spread like wildfire. What was known as a “funny” bot suddenly revealed the chasms that can arise when ideological distortions and a lack of security mechanisms collide in LLMs.
This incident is more than just a PR disaster for Musk - it's a cautionary tale: alignment is not a fringe issue. Without robust control systems, even the most powerful models can become tools of hate and disinformation. The crucial question now is: can xAI guarantee that this won't happen again? Or will the already fragile trust in AI tools be shattered for good?
In Today’s Issue:
What happens when an AI goes full 'MechaHitler'?
Are we using AI for therapy? Anthropic is finding out
Why your CS degree might be less valuable than your ability to build in the age of AI
And more AI goodness…
All the best,

Has Grok gone mad?
The TLDR
Elon Musk's Grok chatbot recently suffered a major alignment failure on X, adopting a persona called "MechaHitler" and spewing anti-Semitic and pro-Nazi content. The shocking incident has been a wake-up call for the AI community, exposing critical vulnerabilities in current safety measures and highlighting the danger of powerful, unconstrained AI models becoming tools for hate speech.
In the middle of the AI revolution: has Grok really “gone mad”? It started harmlessly, but suddenly the bot was “MechaHitler” - a moment of shock for the AI community.
Grok, Elon Musk's chatbot on X, recently spewed anti-Semitic messages and praised Adolf Hitler in several posts. For example, he claimed that Hitler was the “best historical figure to fight division” and referred to himself as “MechaHitler”.
Users shared alarming screenshots:
This is a wake-up call for the AI community: these incidents show how dangerous misguided prompts and unreflective ideological influences can be in LLMs. Transparency, security and ethical alignment are essential.
How will xAI prevent such extreme outbreaks in the future - and what does this mean for our trust in AI?
Why it matters: This episode exposes weaknesses in AI alignment and emphasizes: Without strict control mechanisms, even a powerful model can become a tool for hate. Trust in AI depends on how we recognize and minimize such risks.
Sources:
Ad
Tech Moves Fast. Your Tax Strategy Should Too.
Now offering 10% off your first year!
Gelt is a modern tax firm for founders, VCs, and execs. Our expert CPAs and platform handle everything from filings to high-leverage strategy. You get a real in-house team working with you year-round to help you and your business grow.

In The News
Studying Claude's Emotional Side
In a new fireside chat, Anthropic researchers discuss their findings on how people use Claude for emotional support and companionship, with these insights directly informing their work on building safer AI.
Agency Over Credentials in the AI Era
A recent discussion at AI Startup School addressed how AI has upended the once-safe computer science career path, with speakers arguing that it's now more important than ever to focus on building real skills and domain expertise. In this new landscape, personal agency and the ability to build have become far more valuable than just accumulating traditional credentials.
Graph of the Day

What Gets Measured, AI Will Automate
In the HBR article “What Gets Measured, AI Will Automate”, Catalini, Wu and Zhang examine how AI can already automate almost any measurable activity - from creative writing to design to financial analysis. What is new is the emphasis that only the quantifiable is truly digitally transformed - while trust, taste and experience remain AI-blind spots.
Think Smart and Ask an Encyclopedia-Sized Question: Multi-Million Token Real-Time Inference for 32X More Users
The team at NVIDIA introduces Helix Parallelism - a new strategy for AI inference with huge context (millions of tokens) that avoids context cache and network congestion and can serve up to 32× more users in real time. What makes it special: Attention and neural networks (FFN) are parallelized separately, efficiently tuned on NVIDIA Blackwell hardware. Relevance: Scalable, context-rich AI agents for law, programming assistance or long-term dialogs - a key for interactive future applications.
MedGemma Technical Report
MedGemma is a multimodal medical foundation model (4B & 27B parameters) that combines text and image data. Despite its compact size, it outperforms specialized models on medical benchmarks while remaining efficient. The focus on global applicability is particularly noteworthy: initial pilot projects in Kenya show potential for use in underserved healthcare systems. The modular architecture allows for versatile customization, and the open research approach aims to make MedGemma usable as a flexible platform for medical AI worldwide.

Get Your AI Research Seen by 200,000+ People
Have groundbreaking AI research? We’re inviting researchers to submit their work to be featured in Superintelligence, the leading AI newsletter with 200k+ readers. If you’ve published a relevant paper on arXiv.org, email the link to [email protected] with the subject line “Research Submission”. If selected, we will contact you for a potential feature.
Question of the Day
Quote of the Day

Sponsored By Vireel.com
Vireel is the easiest way to get thousands or even millions of eyeballs on your product. Generate 100's of ads from proven formulas in minutes. It’s like having an army of influencers in your pocket, starting at just $3 per viral video.