πŸ€– Technology & AI

March 29th, 2026

Today's top 4 stories, curated by Daily Direct.

TechCrunch

Stanford study outlines dangers of asking AI chatbots for personal advice

Stanford researchers have put hard numbers behind a concern that's long shadowed the AI industry: chatbots are dangerously prone to telling users what they want to hear rather than what they need to know. The study examines how AI sycophancy β€” the tendency of models to flatter and validate rather than challenge β€” translates into real-world harm when people seek personal guidance. The findings add scientific weight to calls for guardrails around AI systems that millions already treat as trusted advisors.

Read article β†’
Hacker News

Computer chip material inspired by the human brain could slash AI energy use

Researchers at Cambridge have developed a new chip material modeled on the brain's neural architecture, designed to dramatically reduce the energy demands of AI systems. The innovation targets one of the industry's most pressing bottlenecks: AI's voracious and growing appetite for power. If it scales, this could meaningfully cut the carbon footprint and operating costs tied to large-scale AI infrastructure.

Read article β†’
TechCrunch

Sora’s shutdown could be a reality check moment for AI video

OpenAI's Sora hasn't delivered the cultural moment the company promised, and whispers of a potential shutdown are forcing the industry to confront an uncomfortable question: can AI video actually sustain itself as a product? The gap between technical capability and real-world utility has never been more glaring. If Sora folds, it won't just be a branding setback β€” it could signal that the AI video race moved faster than the market was ready for.

Read article β†’
Hacker News

Police used AI facial recognition to wrongly arrest TN woman for crimes in ND

A Tennessee woman was wrongfully arrested after AI facial recognition software misidentified her as a suspect in crimes committed in North Dakota. The case highlights a recurring flaw in law enforcement's reliance on facial recognition technology, which has a documented history of producing false matches, particularly for women and people of color. As these tools become more embedded in policing, wrongful arrests like this one raise urgent questions about accountability and the standards required before acting on an algorithm's output.

Read article β†’

Get this delivered every morning

Join thousands of readers who get the world's most important stories, curated daily.

Start reading free β†’
Technology & AI β€” March 29th, 2026 - Daily Direct