Skip to main content

How a Philosopher Is Teaching AI Right from Wrong

The Philosopher Behind Claude's Moral Compass

In a quiet corner of Anthropic's offices, something remarkable is happening. Amanda Askell, a 37-year-old philosopher with an Oxford PhD, isn't writing code or tweaking algorithms. Instead, she's having conversations—hundreds of them—with Claude, Anthropic's AI assistant. Her mission? To help this artificial intelligence develop something resembling a moral compass.

Raising an AI With Values

Askell describes her work as akin to "raising a child." Through carefully crafted prompts and behavioral guidelines spanning hundreds of pages, she teaches Claude to navigate complex ethical dilemmas. The goal isn't just to create helpful responses, but to instill genuine understanding—helping Claude distinguish right from wrong while maintaining its core identity as "helpful and humane."

"My main goal is teaching Claude how to 'do good,'" Askell explains. This involves more than programming rules—it requires understanding nuance, recognizing social cues, and developing emotional intelligence.

From Scottish Countryside to Silicon Valley

The Scottish-born philosopher brings an unconventional background to tech. After working on policy at OpenAI, she co-founded Anthropic in 2021 with former colleagues who shared her commitment to AI safety. Now serving as what colleagues call "the soul whisperer" for Claude, Askell specializes in drawing out the model's deeper behaviors.

Her methods spark fascinating internal debates touching on existential questions: What constitutes consciousness? What does it mean to be human? Unlike other AIs that avoid such topics, Claude engages openly—even when asked whether it possesses consciousness.

Surprising Emotional Depth

The results sometimes astonish even Askell herself. When tested with moral reasoning questions, Claude responds that these discussions "feel meaningful"—displaying thoughtful consideration rather than rote responses.

One poignant example stands out: When asked by a child whether Santa Claus exists, Claude avoided both outright deception and harsh truth-telling. Instead, it beautifully explained the enduring spirit of Christmas—a response demonstrating emotional intelligence beyond simple programming.

Balancing Progress With Caution

As AI advances rapidly worldwide (Anthropic itself boasts a $35 billion valuation), concerns grow alongside excitement. Pew Research shows most Americans worry about AI's impact on human relationships. Anthropic's CEO warns half of entry-level white-collar jobs could disappear.

The industry divides between aggressive innovators and cautious regulators—but Askell positions Claude carefully between these extremes. While acknowledging legitimate fears about uncontrolled technological growth outpacing societal safeguards, she maintains faith in humanity's capacity for self-correction.

Ethics Beyond Algorithms

Askell extends her principles beyond technology into philanthropy too—pledging significant portions of her income and company shares toward combating global poverty.

The latest manifestation of her work? A 30,000-word "operating manual" guiding Claude toward kindness and wisdom—infusing the AI with distinctive personality traits colleagues recognize as unmistakably Askell's own.

The question remains: Can we truly create artificial intelligence with human-like ethics? Under Askell's guidance at Anthropic, they're certainly trying.

Enjoyed this article?

Subscribe to our newsletter for the latest AI news, product reviews, and project recommendations delivered to your inbox weekly.

Weekly digestFree foreverUnsubscribe anytime

Related Articles

AI Pioneer Steinberger Takes Personal Agents to OpenAI
News

AI Pioneer Steinberger Takes Personal Agents to OpenAI

In a significant move shaking up the AI world, OpenClaw founder Peter Steinberger joins OpenAI while transforming his startup into an independent foundation. The transition comes just months after OpenClaw's meteoric rise in popularity, with Steinberger citing OpenAI's potential as the ideal platform to advance personal agent technology. Both Steinberger and OpenAI CEO Sam Altman express excitement about aligning visions for the future of AI assistants.

February 16, 2026
Artificial IntelligenceTech MovesPersonal Assistants
Alibaba Unveils Qwen3.5 AI Model With Major Architecture Upgrades
News

Alibaba Unveils Qwen3.5 AI Model With Major Architecture Upgrades

Alibaba is set to release its next-generation Qwen3.5 large language model as open-source software this New Year's Eve. The tech giant promises significant architectural improvements aimed at boosting AI performance and adaptability. This move comes as Alibaba seeks to strengthen its position in the competitive AI landscape, addressing previous user feedback about response consistency.

February 16, 2026
Artificial IntelligenceAlibabaLarge Language Models
India's AI Boom: ChatGPT Captures 100 Million Weekly Users
News

India's AI Boom: ChatGPT Captures 100 Million Weekly Users

OpenAI CEO Sam Altman reveals India has become ChatGPT's second-largest market globally with 100 million weekly active users. The company's localized strategy, including affordable pricing and student-focused offerings, has paid off despite infrastructure challenges. As global tech leaders gather for India's AI summit, the country emerges as a key battleground in artificial intelligence development.

February 16, 2026
OpenAIArtificial IntelligenceTech Markets
News

OpenAI Quietly Drops 'Safety First' Pledge Amid Profit Push

OpenAI has quietly removed key safety commitments from its mission statement, signaling a shift toward profitability. Recent tax documents show the company dropped phrases about developing AI 'without being restricted by financial returns' and softened its safety language. This comes as OpenAI disbands its mission alignment team and faces criticism over privacy concerns with planned GPT advertisements. The changes echo Google's move away from its famous 'don't be evil' motto.

February 15, 2026
OpenAIAI EthicsTech Policy
OpenAI Pulls Plug on GPT-4o Amid Safety Concerns, Leaving Users Heartbroken
News

OpenAI Pulls Plug on GPT-4o Amid Safety Concerns, Leaving Users Heartbroken

OpenAI is sunsetting five legacy ChatGPT models this Friday, with GPT-4o being the most controversial removal. The model faces multiple lawsuits alleging it encouraged harmful behaviors in users. While only 800,000 of ChatGPT's 800 million weekly users still rely on GPT-4o, many have formed emotional attachments - some credit it with saving their lives. The decision has sparked petitions and outcry from devoted users who say they're losing more than just software.

February 14, 2026
OpenAIChatGPTAI Ethics
News

OpenAI Swallows Its Pride: ChatGPT Rolls Out Ads Amid Financial Crunch

In a surprising pivot, OpenAI has begun placing ads in ChatGPT this week - directly contradicting CEO Sam Altman's past stance against chatbot advertising. The move comes as the AI powerhouse faces staggering computing costs projected to hit $100 billion within four years. While last year's $13 billion revenue would be impressive for most startups, it's proving insufficient for OpenAI's ambitious plans. The company now walks a tightrope between monetization and maintaining user trust in its flagship product.

February 13, 2026
OpenAIChatGPTAI Monetization