Dozens of attorneys general across the United States have sent warning letters to major artificial intelligence companies — including OpenAI, Anthropic, Perplexity, and others — demanding explanations regarding system safety, data protection, safeguards against misuse, and the prevention of false or harmful AI-generated content.
The Japanese startup Integral AI has announced that it has created the “world’s first AGI-capable model.” According to a company representative, the team has achieved one of the ultimate goals of artificial intelligence development: enabling their AI model to learn new tasks “without pre-existing datasets or human intervention.”
Large language models (LLMs) that already assist physicians and patients with medical questions can still generate severely harmful advice in a sizable percentage of real cases, according to a new multi-center preprint introducing the NOHARM safety benchmark.
Artificial intelligence is becoming one of the primary tools for emotional support among teenagers. Young users are increasingly turning to chatbots, treating them as a safe, anonymous, and instantly available source of help — especially at a time when access to professional therapy is limited.
