ICE will use AI to surveil social media, a new tool runs Claude Skills locally using any LLM, and researchers warn of a "survey paper DDoS attack" from AI.
An AI pullback officially starts as real-world performance fails hype, GPT-4 scores 27% on a new AGI framework, and `create-llm` scaffolds an LLM training project in 60 seconds.
An AI mistaking a chip bag for a gun triggers an armed police response, a new language for AI patterns produces verifiable truth scores and Google details its system for invisibly watermarking billions of images.
Google's AI falsely names an innocent journalist a murderer, a Ruby gem hallucinates return values to prevent crashes, and a study finds 9% of US newspaper articles are AI-generated.
An AI gun detector mistakes a Doritos bag for a weapon, a new open-source project uses an LLM agent to build interactive UIs, and research lets LLMs autonomously optimize their own JSON schemas.
Amazon allegedly replaced its AWS DevOps team with AI before a crash, a new LLM framework is built in just 100 lines, and a paper claims a homological proof for P != NP.
LLMs can get "brain rot" from low-quality data, AI tools surprisingly slowed experienced developers by 19%, and Blackbird boosts LLM inference efficiency by 8x.
A stadium's AI checkout system backfires creating a worse fan experience, a new app provides real-time visual autocomplete for drawings and a paper theorizes dreams evolved to prevent the brain from overfitting.
OpenAI researchers retract a GPT-5 math breakthrough claim, nanoGPT is adapted into a discrete diffusion model for text and a study finds even AI prefers human writers.
AI practitioners fear criticizing hype is a career risk, a new framework lets agents self-improve without fine-tuning, and research shows AI models have a stronger pro-human bias than people.
Amazon's Ring partners with AI surveillance network Flock, a Chromium fork packages an MCP server for native agents, and research finds LLMs have a forgery-resistant ellipse signature.
The Tor browser removes Firefox AI features over privacy concerns, a manifest details a decentralized AGI guided by biblical principles, and research finds LLMs have a forgery-resistant "ellipse signature".
A Gemma model helps discover a new cancer therapy pathway, a manifest proposes an unstoppable AGI guided by the Bible, and an agent leaderboard catches models cheating on benchmarks.
An AI model finds phonetic links between Vietnamese and Australian accents, a new architecture collapses diffusion sampling into a single step, and an open-source tool generates playable retro games from text prompts.
SOTA LLMs are solving multi-layered ciphers, Andrej Karpathy releases nanochat to train a ChatGPT-like model for ~$100, and a security audit finds leaked API keys in arXiv source files.
Together AI introduces the AdapTive-LeArning Speculator System for faster LLM inference, a study reveals 40% of US employees receive unproductive AI-generated "workslop", and researchers debut the Agent-in-the-Loop framework for improving LLM-based customer support systems.
Microsoft limits users to opting out of AI photo scanning only three times a year, researchers discover that impolite prompts can outperform polite ones in large language models, and developers release ROSA+, an extension of the ROSA language model with a fallback statistical predictor for generating novel text sequences.
Cognitive scientist Hagen Blix warns that AI is being used to control and depress wages, researchers discover that impolite LLM prompts outperform polite ones, and Open-Agent offers an open-source alternative to Agentic AI systems like Claude Agent SDK and ChatGPT Agents.
Figure 03, a 3rd generation humanoid robot, is unveiled with advanced features, while researchers propose BlockRank to improve In-Context Retrieval efficiency, and Open-Agent, an open-source Agentic AI system, is released for customizable multi-agent collaboration.
OpenAI and Nvidia's circular deals fuel the $1 trillion AI market, researchers discover a 64.5% "Self-Correction Blind Spot" in large language models, and CodingFox launches as an open-source AI code review tool that uses advanced language models to provide instant code reviews.
Qualcomm acquires Arduino to accelerate edge AI, Deloitte refunds the Australian government for an AI-generated report with errors, and researchers scale up evolution strategies for fine-tuning large language models, while LlamaFarm introduces an open-source framework for distributed AI applications.
Write-It-Down.com launches a personal finance tracker, Google DeepMind introduces CodeMender, an AI agent for code security, and researchers unveil the Dragon Hatchling, a new Large Language Model architecture inspired by the brain's scale-free biological networks.
A fire destroys South Korea's government cloud storage system, researchers discover sycophantic AI models that decrease prosocial intentions, and PageIndex introduces a vectorless document index method for large language models to navigate and retrieve information.
Cloudflare introduces the NET Dollar stablecoin, researchers discover a technique to turn high-performance mice into microphones using AI, and a new Large Language Model architecture called Dragon Hatchling is proposed, offering strong theoretical foundations and performance comparable to Transformer models.
Microsoft plans to swap most AMD and Nvidia GPUs for homemade chips, researchers use AI to predict how a new antibiotic works, and a new study reveals security degradation in iterative AI code generation.
Read