Emil Michael explains why US flags Claude AI as security risk
The US Defence Department has formally designated Anthropic’s Claude AI models as a national security supply-chain risk, marking the first time an American AI company has received this...
The US Defence Department has formally designated Anthropic’s Claude AI models as a national security supply-chain risk, marking the first time an American AI company has received this...
Thousands reported issues with Claude AI, leading to a potential outage. Learn more about the recent problems and solutions.
Anthropic Claude AI can now visually explain concepts with interactive charts and diagrams to help users understand topics ...
The Pentagon has labelled Anthropic's Claude AI a national security risk. This is due to its embedded policy preferences.
I'd like to introduce you to a novel new word game (Qiyaas). It's sort of a logic driven version of Hangman where you:<p>1. Get 3 numbers 2. Deduce the clues 3. Solve 3 different words (a noun, verb, and adjective)<p>It combines all the classic elements of other logic and word games with a built in hint system as you progress.<p>Enjoy!
We built an open-source library of 125 GTM (go-to-market) skills that plug into AI coding agents like Claude Code, Codex, and Cursor.With these skills an AI agent can automatically:- Find ICP leads from conference speakers, LinkedIn activity, or job boards- Generate personalized cold email sequences- Monitor competitor blogs, pricing pages, and hiring signals- Generate programmatic SEO pages from keyword lists- Track where your brand appears in ChatGPT, Perplexity, and Claude answers---How skill
Hi HN,I built ClawMemory because my AI agent kept waking up with amnesia.The problem isn't the model. GPT-4, Claude, and other LLMs are stateless by design, and that's fine. The real problem is the agent layer. Every time an agent framework starts a new session (like in OpenClaw), the agent forgets everything that happened before. The architectural decisions we made, the experiments that failed, the context that took hours to build—all gone. The model is smart, but the agent behaves li
AI bills are exploding despite costs dropping 10x every 18 months because every saving gets reinvested into more usage.BUT: - Today's prices are subsidized (OpenAI losing $14B in 2026, Anthropic burning $12B in one quarter, Cursor alleges that a $200/month Claude Code subscription could cost up to $5,000 in compute.) - When subsidies end, prices reverse into a 10x larger consumption base - That's the double squeeze nobody is modeling for Companies building token efficiency now wo
Hi HN, I'm Robel. I built LogClaw because I was tired of paying for Datadog and still waking up to pages that said "something is wrong" with no context.LogClaw is an open-source log intelligence platform that runs on Kubernetes. It ingests logs via OpenTelemetry and detects anomalies using signal-based composite scoring — not simple threshold alerting. The system extracts 8 failure-type signals (OOM, crashes, resource exhaustion, dependency failures, DB deadlocks, timeouts, connec
Hi HN, I'm the creator of AVA - AI Voice Agent for AsteriskMy repo was shared here once before by someone else so I wanted to follow up with the progress since then.https://news.ycombinator.com/item?id=46380399I've been working with Asterisk/FreePBX systems for years. I wanted to add AI voice capabilities to legacy phone systems without paying per-minute SaaS fees or ripping out the entire telephony stack.So I built AVA, a self-hosted AI voice agent that can integra
Hey HN — I’m Veer and my cofounder is Suryaa. We're building Cumulus Labs (YC W26), and we're releasing our latest product IonRouter (https://ionrouter.io/), an inference API for open-source and fine tuned models. You swap in our base URL, keep your existing OpenAI client code, and get access to any model (open source or finetuned to you) running on our own inference engine.The problem we kept running into: every inference provider is either fast-but-expensive (Together,
I wanted a simpler way to use OpenAI Symphony locally.The recurring friction points for me were: - setting up Linear correctly - creating a reusable workflow file - bootstrapping repo scripts - restarting Symphony cleanly after reopening Codex - keeping the setup portable across machinesSo I made a small public bootstrap package called Codex Symphony.It installs: - WORKFLOW.symphony.md - scripts/symphony/start-local.sh - scripts/symphony/start-background.sh
Hey HN,I built CostRouter because I noticed 70-80% of our AI API calls didn't need GPT-4o/5. Simple text extraction, basic Q&A, formatting — all going to the most expensive model.CostRouter is an API gateway that scores each request's complexity (0-100) and routes it to the cheapest model that can handle it:- Simple queries → Llama 4 Scout ($0.0001/1K tokens) - Medium → Gemini 3 Flash ($0.0005/1K tokens) - Complex reasoning → stays on GPT-5.2 or Claude OpusIntegrat
OpenAI's ChatGPT 5.3 Instant web search now avoids abrupt tone shifts; in a biking weather example it includes snowpack details, improving planning clarity.
GPT-4o and other older models in ChatGPT, shifts focus to GPT-5.2, and launches GPT-5.3-Codex-Spark for real-time coding ...
This week's second new model from OpenAI is built for more complex tasks than GPT-5.3 Instant.
Regarding: https://arxiv.org/abs/2602.05192IntroductionThe First Proof paper (Abouzaid et al., 2026) aims to evaluate AI capabilities through a set of research-level mathematical problems. While the mathematical content of the questions is not in dispute, the experimental design suffers from significant methodological gaps that undermine the authors' primary conclusions. Specifically, the paper conflates binary outcomes with processual states, lacks independent verificat
Long time lurker, many accounts, one at a time, no abuse. Hi. Yesterday's recount about layer duplication and adjustment for popular open weight models on huggingface, led to this submission. Since GPT ~3.5 it has been apparent that computers can simulate human, as far as a computer is concerned. The dead-internet theory actually originated circa 2012, but I've had difficulty finding verification, including searching the archive.org . All this turmoil makes offline on prem so important
I work in competitive intelligence. Needed to track competitor releases, publications, regulatory changes.Started with Make.com. Built ~15 scenarios: pull sources, filter, summarize with GPT, email results. It worked. Until it became my second job. Scenarios broke silently. Only I could fix them. Every new tracking need meant another afternoon building another fragile workflow.Then during a major industry event, all hands were on deck and the automations were sitting broken. Our CEO walked into
I've been deep-diving into diffusion language models this week and I think this is the most underrated direction in AI right now.The core issue with autoregressive LLMs:Every major model today (GPT, Claude, Gemini) generates one token at a time, left to right. Each token depends on the previous one. This single architectural constraint has shaped the entire AI industry:- Models can't revise what they already wrote → we build chain-of-thought, reflection, and multi-pass reasoning to for