GPT-5

Explained: What is behind the Pentagon’s clash with Anthropic?

Pentagon warns Anthropic over military use of its AI model. Dispute centres on safeguards around surveillance and autonomous ...

Anthropic Leans Into Enterprise With Managed Claude Cowork Plugins

Anthropic this week announced a new plugin ecosystem and extensions to the Cowork platform that makes it easier for enterprises to build and manage workflow integrated agents.

Pentagon threatens to make Anthropic a pariah if it refuses to drop AI guardrails

Defense Secretary Pete Hegseth gave Anthropic CEO Dario Amodei a Friday deadline to comply with demands to peel back ...

Anthropic says Claude Code transformed programming. Now Claude Cowork is coming for the rest of the enterprise.

Anthropic is making its boldest enterprise push yet with Claude Cowork, rolling out private plug-in marketplaces, deep integrations, and AI agent tools that are reshaping corporate adoption and ...

Anthropic’s New Claude Plug-Ins Take Aim at Finance, HR, and More. Is Your Job Next?

Anthropic is disrupting knowledge work with new plug-ins capable of financial analysis, investment banking, engineering, and much more.

Pentagon threatens to cancel Anthropic's $200 million contract over AI safeguards

Defense Secretary Pete Hegseth has given Anthropic until Friday to ease safety restrictions on its AI model, Claude, or risk ...

Anthropic ditches its core safety promise in the middle of an AI red line fight with the Pentagon

Anthropic, a company founded by OpenAI exiles worried about the dangers of AI, is loosening its core safety principle in ...

Hegseth gives Anthropic deadline to give military full access to its AI

Defense Secretary Pete Hegeseth has given Anthropic until Friday to give the U.S. military unfettered access to its AI model, ...

Samsung Galaxy S26's Perplexity AI Integration is Deeper Than You Think

The post Samsung Galaxy S26's Perplexity AI Integration is Deeper Than You Think appeared first on Android Headlines.

GPT-5 bests human judges in legal smack down

But that doesn't mean AI is ready to dispense justice ai-pocalypse Legal scholars have found that OpenAI's GPT-5 follows the ...

OpenAI's latest GPT-5.3-Codex and audio models now on Microsoft Foundry

OpenAI has expanded the availability of its GPT-5.3-Codex model to third-party developers via API and Microsoft Foundry.

Say please? The best way to talk to an AI

"Let's say you want to generate a job description. Tell the AI 'I want you to ask me questions, one at a time, until you've ...

AI nerves are fraying. Anthropic keeps doubling down

Just weeks after its AI tools shook software stocks, Anthropic is pushing even deeper into the workplace. The company is ...

Claude the conqueror: The AI chatbot keeps wiping out billions from exposed tech stocks

Claude, Anthropic's AI chatbot, has sparked major selling of US tech stocks as its capabilities induce fear among investors ...

What smart people in economics and business are saying about a viral report warning of an AI-driven recession and stock crash

A worst-case AI scenario rattled markets, drawing sharp pushback from economists and business leaders.

Medical AI Is Already In Hospitals. Who Is Watching Its Safety?

The FDA’s oversight was built for devices that rarely change. Clinical AI evolves over time, raising new questions about who ...

I spent $100 benchmarking LLM providers on a weekend CTF

This past weekend, I decided to test out a cli tool I've been building to help me do source code reviews _faster_.I figured the best environment for such a tool would be a Weekend CTF event. I like web challenges since you get a nice dump of source code, as well as a Dockerfile or docker compose setup for how to run everything locally. Usually, I can complete 2-3 Web challenges before I get stuck. To help get unstuck I found myself increasingly turning to LLMs as a pairing partner.I'm

Show HN: Open-source EU AI Act compliance layer for AI agents (8/2026 deadline)

We built AIR Blackbox — open-source compliance infrastructure for AI agents targeting the EU AI Act enforcement deadline on August 2, 2026. If you're deploying LLM-based agents (LangChain, CrewAI, AutoGen, OpenAI Agents SDK) into production, the EU AI Act requires tamper-evident audit trails, human oversight mechanisms, data governance controls, and injection defense — for any system classified as high-risk. Most teams we've talked to either don't know about the deadline or assume

Terms of use: What types of competition do model providers ban?

I thought it would be interesting to look at the terms of service of the frontier labs and there was more deviation than I expected when it comes to the issue of building competing offerings. Note that I am not a lawyer and none of this is legal advice. You should refer to the specific versions of the agreements that apply to you and consult with a lawyer.It is very common for technology companies (particularly when providing data through an API,) to include a term that more-or-less says their c

Show HN: I applied Markowitz port. theory to agent teams / proved it in a zkVM

I run multi-agent teams in high-consequence scenarios. Read: fuckups at 3 AM = I'm awake.I kept hitting the same issue. I couldn't get a rules-based system to enforce behavior and I had no real way to prove that agents really did what they said they did. I can log and monitor them - set up (a million) Slack alerts but none of these things are PROOF. Logs are mutable. And that matters more every day as agents get more powerful (take THAT, @meta)So I went down the rabbit hole.The obvious