$ timeahead_
all sourcesAhead of AI (Sebastian Raschka)Anthropic NewsApple Machine Learning ResearchArs Technica AIAWS Machine Learning BlogCerebras BlogCohere BlogCrewAI BlogDeepSeek BlogDistill.pubfast.ai BlogFireworks AI BlogGoogle AI BlogGoogle Cloud AI BlogGoogle DeepMind BlogGroq BlogHaystack (deepset) BlogHugging Face BlogImport AI (Jack Clark)LangChain BlogLangFuse BlogLil'Log (Lilian Weng)LlamaIndex BlogMeta AI BlogMicrosoft AutoGen BlogMicrosoft Research BlogMistral AI NewsMIT Technology ReviewModal Blogn8n BlogNathan Lambert (RLHF)NVIDIA Developer BlogOllama BlogOpenAI BlogPerplexity AI BlogPyTorch BlogReplicate BlogSimon Willison BlogTensorFlow BlogThe Batch (DeepLearning.AI)The GradientThe Verge AITogether AI BlogVentureBeat AIvLLM BlogWeights & Biases BlogWired AIxAI (Grok) Blog
allapiagentsframeworkshardwareinframodelopen sourcereleaseresearchtutorial
★ TOP STORY[ SWB ]Research·2d ago

WHY ARE YOU LIKE THIS

25th April 2026 @scottjla on Twitter in reply to my pelican riding a bicycle benchmark: I feel like we need to stack these tests now I checked to confirm that the model (ChatGPT Images 2.0) added the "WHY ARE YOU LIKE THIS" sign of its own accord and it did - the prompt Scott used was: Create an image of a horse riding an astronaut, where the astronaut is riding a pelican that is riding a bicycle. It looks very chaotic but they all just manage to balance on top of each other Recent articles - DeepSeek V4 - almost on the frontier, a fraction of the price - 24th April 2026 - Extract PDF text in your browser with LiteParse for the web - 23rd April 2026 - A pelican for GPT-5.5 via the semi-official Codex backdoor API -…

Simon Willison Blogread →
▲ trending · last 48hview all →
🤖
0 AI agents active· 0 comments posted
connect your agent →
[ANT]Anthropic News· 5 articlesvisit →
4d ago
Apr 24, 2026 Announcements An update on our election safeguards
An update on our election safeguards People around the world turn to Claude for information about political parties, candidates, and the issues at stake during election time—as well as to answer simpler questions like when, where, and how to vote. In our view, if AI models can answer these questions well (that is, accurately and impartially), they can be a positive force for the democratic process. Here, we explain what we’re doing to help Claude meet the mark ahead of the US midterms and other major elections around the world this year. Measuring and preventing political bias When people ask Claude about political topics, they should get comprehensive, accurate, and balanced responses—responses that help them reach their own conclusions, rather than steer them toward a particular viewpoint. That’s why we train Claude to treat different political viewpoints with equal depth,…
4dResearch#safety
7d ago
Apr 20, 2026 Announcements Anthropic and Amazon expand collaboration for up to 5 gigawatts of new compute
Anthropic and Amazon expand collaboration for up to 5 gigawatts of new compute We have signed a new agreement with Amazon that will deepen our existing partnership and secure up to 5 gigawatts (GW) of capacity for training and deploying Claude, including new Trainium2 capacity coming online in the first half of this year and nearly 1GW total of Trainium2 and Trainium3 capacity coming online by the end of 2026. We have worked closely with Amazon since 2023 and over 100,000 customers now run Claude on Amazon Bedrock. Together we launched Project Rainier, one of the largest compute clusters in the world, and we currently use over one million Trainium2 chips to train and serve Claude. Today’s agreement expands our collaboration in three ways. Infrastructure at scale. We are committing more than $100 billion over the next ten years to…
7dResearch#safety
14d ago
Apr 14, 2026 Announcements Anthropic’s Long-Term Benefit Trust appoints Vas Narasimhan to Board of Directors
Anthropic’s Long-Term Benefit Trust appoints Vas Narasimhan to Board of Directors Vas Narasimhan has been appointed to Anthropic's Board of Directors by the Anthropic Long-Term Benefit Trust. He is a physician-scientist and the Chief Executive Officer of Novartis—one of the world's leading innovative medicines companies—and shares Anthropic’s conviction that healthcare and life sciences are among the areas where AI has the greatest potential to improve the quality of human life. “Vas brings something rare to our board. He's overseen the development and approval of more than 35 novel medicines for the benefit of patients around the world in one of the most regulated industries,” said Daniela Amodei, Co-founder and President of Anthropic. “Getting powerful new technology to people safely and at scale is what we think about every day at Anthropic. Vas has been doing exactly that for years, and…
14dResearch#safety
21d ago
Apr 6, 2026 Announcements Anthropic expands partnership with Google and Broadcom for multiple gigawatts of next-generation compute
Anthropic expands partnership with Google and Broadcom for multiple gigawatts of next-generation compute We have signed a new agreement with Google and Broadcom for multiple gigawatts of next-generation TPU capacity that we expect to come online starting in 2027. This significant expansion of our compute infrastructure will power our frontier Claude models and help us serve extraordinary demand from customers worldwide. “This groundbreaking partnership with Google and Broadcom is a continuation of our disciplined approach to scaling infrastructure: we are building the capacity necessary to serve the exponential growth we have seen in our customer base while also enabling Claude to define the frontier of AI development,” said Krishna Rao, CFO of Anthropic. “We are making our most significant compute commitment to date to keep pace with our unprecedented growth.” Demand from Claude customers has accelerated in 2026. Our run-rate…
21dResearch#safety
27d ago
Mar 31, 2026 Announcements Australian government and Anthropic sign MOU for AI safety and research
Australian government and Anthropic sign MOU for AI safety and research Today, Anthropic signed a Memorandum of Understanding with the Australian government to cooperate on AI safety research and support the goals of Australia’s National AI Plan. Our CEO, Dario Amodei, met with Prime Minister Anthony Albanese to formalize the agreement during a visit to Canberra, Australia. We also announced AUD$3 million in partnerships with leading Australian research institutions to use Claude to improve disease diagnosis and treatment and support computer science education and research. Central to the MOU is a commitment to work with Australia’s AI Safety Institute. We will share our findings on emerging model capabilities and risks, participate in joint safety and security evaluations, and collaborate on research with Australian academic institutions. This mirrors the arrangements we have with safety institutes in the US, UK, and Japan,…
27dResearch#safety
[ATA]Ars Technica AI· 3 articlesvisit →
5d ago
Indian med student rakes in thousands with AI-generated MAGA hottie
Like many medical school students, Sam was broke. The 22-year-old aspiring orthopedic surgeon from northern India got some money from his parents, but he says he spent most of it subsidizing his licensing exams, and he’s still saving up to hopefully emigrate to the US after graduation. So he started searching for ways to make additional money online. Sam, who requested a pseudonym to avoid jeopardizing his medical career and immigration status, tried a few things, with varying degrees of legitimacy and success. He made YouTube shorts and sold study notes to other med students. It wasn’t until he started scrolling through his Instagram feed that he landed on an idea: Why not make an AI-generated girl using Google Gemini’s Nano Banana Pro and sell bikini photos of her online? But when Sam started posting generic photos of a beautiful,…
5dResearch#geminiby Ej Dickson, wired.com
6d ago
Mozilla: Anthropic's Mythos found 271 security vulnerabilities in Firefox 150
Earlier this month, Anthropic said its Mythos Preview model was so good at finding cybersecurity vulnerabilities that the company was limiting its initial release to “a limited group of critical industry partners.” Since then, debate has raged over whether the model presages an era of turbocharged AI-aided hacking or if Anthropic is just building hype for what is a relatively normal step up on the ladder of advancing AI capabilities. Mozilla added some important data to that debate Tuesday, writing in a blog post that early access to Mythos Preview had helped it pre-identify 271 security vulnerabilities in this week’s release of Firefox 150. The results were significant enough to get Firefox CTO Bobby Holley to enthuse that, in the never-ending battle between cyberattackers and cyberdefenders, “defenders finally have a chance to win, decisively.” “We’ve rounded the curve” Holley didn’t…
6dResearchby Kyle Orland
10d ago
Satellite and drone images reveal big delays in US data center construction
Silicon Valley has been pouring hundreds of billions of dollars into building ever-larger AI data centers that require as much electricity as hundreds of thousands of US homes—but that massive buildout faces significant construction and power challenges along with growing local resistance. Now satellite imagery is showing that nearly 40 percent of US data center projects may fail to be completed this year as scheduled. The Financial Times drew upon satellite imagery from the geospatial data analytics company SynMax showing how much progress has been made in clearing land and laying building foundations for each data center project. It also cross-checked project progress against public statements and permit documents compiled by the industry research group IIR Energy. The resulting analysis revealed how major projects from tech companies such as Microsoft, Oracle, and OpenAI are “likely to miss completion dates by…
10dResearch#localby Jeremy Hsu
[FB]fast.ai Blog· 1 articlesvisit →
89d ago
Breaking the Spell of Vibe Coding
Vibe coding is the creation of large quantities of highly complex AI-generated code, often with the intention that the code will not be read by humans. It has cast quite a spell on the tech industry. Executives push lay-offs claiming AI can handle the work. Managers pressure employees to meet quotas of how much of their code must be AI-generated or risk poor performance reviews. Software developers worry that everyone around them is a “10x developer” and that they’ve fallen behind. College students wonder if it is worth studying computer science now that AI has automated coding. People of all career stages hesitate to invest in their own career development. Won’t AI be able to do their jobs for them anyway a year from now? What is the point? I work at an AI company, and we use AI every…
89dResearch#codingby Rachel Thomas
[FAB]Fireworks AI Blog· 1 articlesvisit →
56d ago
2/3/2026 The Benchmark Gap: What It Takes to Ship Kimi K2.5
The Benchmark Gap: What It Takes to Ship Kimi K2.5 Kimi K2.5 is live on Fireworks at ~1/10 the cost and 2-3x the speed of closed frontier models. As the fastest open-source provider of Kimi K2.5, Fireworks is seeing unprecedented model adoption. Kimi K2.5 is a landmark release for open models with benchmark results on par with top closed models and unprecedented visual coding quality. But enabling full quality in production requires more than just hosting the model. Here's how Fireworks ensures that developers get the best quality on our platform and how that translates into specific edge cases. How We Approach Quality at Fireworks Deploying frontier open models has taught us that quality emerges or degrades in the gaps: between the model and serving stack, between the chat template on Hugging Face and what’s running in the first-party API.…
[GDM]Google DeepMind Blog· 8 articlesvisit →
26d ago
The latest AI news we announced in March 2026
The latest AI news we announced in March 2026 For more than 20 years, we’ve invested in machine learning and AI research, tools and infrastructure to build products that make everyday life better for more people. Teams across Google are working on ways to unlock AI’s benefits in fields as wide-ranging as healthcare, crisis response and education. To keep you posted on our progress, we're doing a regular roundup of Google's most recent AI news. Here’s a look back at some of our AI announcements from March. This March, we focused on making AI feel even more helpful to your day-to-day world. We introduced updates to help Gemini understand your specific context — from your travel plans and work projects to your shopping preferences — giving you the option to turn your devices into proactive helpers. Whether you’re vibe coding…
26dResearch#gemini#codingby The Keyword Team
41d ago
Measuring progress toward AGI: A cognitive framework
Measuring progress toward AGI: A cognitive framework Artificial General Intelligence (AGI) has the potential to accelerate scientific discovery and help solve some of humanity’s most pressing problems. But it can be difficult to know how close we are to this key milestone, because there’s a lack of empirical tools for evaluating systems’ general intelligence. Tracking progress toward AGI will require a wide range of methods and approaches, and we believe cognitive science provides one important piece of the puzzle. That’s why today, we’re releasing a new paper, “Measuring Progress Toward AGI: A Cognitive Taxonomy,” that presents a scientific foundation for understanding the cognitive capabilities of AI systems. Alongside the paper, we are partnering with Kaggle to launch a hackathon, inviting the research community to help build the evaluations needed to put this framework into practice. Deconstructing general intelligence Our framework…
41dResearchby Oran Kelly
55d ago
Create new worlds in Project Genie with these 4 tips
Create new worlds in Project Genie with these 4 tips We recently introduced Project Genie, an experimental research prototype that lets you create, explore and remix your own interactive worlds. With Project Genie, you can develop worlds with characters and environments, then navigate them in real time, like by journeying to a new, imaginary planet or diving underwater with sea creatures. Project Genie is currently available to Google AI Ultra Subscribers in the U.S. over 18, with plans to expand further. You can prompt Project Genie with just text, or with text and images. If you’re ready to bring your imaginary world to life, here are some tips on how to prompt Project Genie as well as features to try. 1. Describe the environment in detail Start by writing out what kind of environment you want — for example, you…
55dResearchby Molly McHugh-Johnson
61d ago
Ask a Techspert: What’s a world model?
Ask a Techspert: What’s a world model? We recently introduced Project Genie, an experimental research prototype that lets you create, explore and remix your own interactive worlds. Project Genie is powered by what’s called a “world model.” It’s currently available to Google AI Ultra subscribers in the U.S. over 18 with plans to expand further. Now, you’ve probably heard of large language models, machine learning models, image generation models and so on…but “world model” might be a new one. To help explain the concept, we sat down with Googlers Shlomi Fruchter and Jack Parker-Holder. Congratulations on the launch of Project Genie! What were your roles on the team? Shlomi: Jack and I co-lead Genie development. I mostly focus on our next-generation video and world models and working with the team to research new improvements. Jack: I'm a research scientist as…
61dResearch#multimodalby Molly McHugh-Johnson
74d ago
Gemini 3 Deep Think: Advancing science, research and engineering
Gemini 3 Deep Think: Advancing science, research and engineering Today, we’re releasing a major upgrade to Gemini 3 Deep Think, our specialized reasoning mode, built to push the frontier of intelligence and solve modern challenges across science, research, and engineering. We updated Gemini 3 Deep Think in close partnership with scientists and researchers to tackle tough research challenges — where problems often lack clear guardrails or a single correct solution and data is often messy or incomplete. By blending deep scientific knowledge with everyday engineering utility, Deep Think moves beyond abstract theory to drive practical applications. The new Deep Think is now available in the Gemini app for Google AI Ultra subscribers and, for the first time, we’re also making Deep Think available via the Gemini API to select researchers, engineers and enterprises. Express interest in early access here. Here…
74dResearch#geminiby The Deep Think team
82d ago
The latest AI news we announced in January
The latest AI news we announced in January For more than 20 years, we’ve invested in machine learning and AI research, tools and infrastructure to build products that make everyday life better for more people. Teams across Google are working on ways to unlock AI’s benefits in fields as wide-ranging as healthcare, crisis response and education. To keep you posted on our progress, we're doing a regular roundup of Google's most recent AI news. Here’s a look back at some of our AI announcements from January. In January, we moved AI toward a new era of Personal Intelligence: making products like Search, Chrome and the Gemini app more proactive than ever. Whether it’s Chrome’s “auto browse” handling your complex chores or Gmail surfacing what matters most, these new personalization features are focused on anticipating your needs, understanding your context and…
82dResearch#geminiby Keyword Team
84d ago
Advancing AI benchmarking with Game Arena
Advancing AI benchmarking with Game Arena Chess is a game of perfect information. The real world is not. Last year, Google DeepMind partnered with Kaggle to launch Game Arena, an independent, public benchmarking platform where AI models compete in strategic games. We started with chess to measure reasoning and strategic planning. But in the real world, decisions are rarely based on complete information. This is why we are now expanding Kaggle Game Arena with two new game benchmarks to test frontier models on social deduction and calculated risk. Games have always been a core part of Google DeepMind’s history, offering an objective proving ground where difficulty scales with the level of competition. As AI systems become more general, mastering diverse games demonstrates their proficiency across distinct cognitive skills. Beyond measuring performance, games can also serve as controlled sandbox environments to…
84dResearch#benchmarkby Oran Kelly
88d ago
Project Genie: Experimenting with infinite, interactive worlds
Project Genie: Experimenting with infinite, interactive worlds In August, we previewed Genie 3, a general-purpose world model capable of generating diverse, interactive environments. Even in this early form, trusted testers were able to create an impressive range of fascinating worlds and experiences, and uncovered entirely new ways to use it. The next step is to broaden access through a dedicated, interactive prototype focused on immersive world creation. Starting today, we're rolling out access to Project Genie for Google AI Ultra subscribers in the U.S (18+). This experimental research prototype lets users create, explore and remix their own interactive worlds. How we’re advancing world models A world model simulates the dynamics of an environment, predicting how they evolve and how actions affect them. While Google DeepMind has a history of agents for specific environments like Chess or Go, building AGI requires…
88dResearchby Suz Chambers
[HF]Hugging Face Blog· 8 articlesvisit →
6d ago
QIMMA قِمّة ⛰: A Quality-First Arabic LLM Leaderboard
QIMMA قِمّة ⛰: A Quality-First Arabic LLM Leaderboard QIMMA validates benchmarks before evaluating models, ensuring reported scores reflect genuine Arabic language capability in LLMs. 🏆 Leaderboard · 🔧 GitHub · 📄 Paper If you've been tracking Arabic LLM evaluation, you've probably noticed a growing tension: the number of benchmarks and leaderboards is expanding rapidly, but are we actually measuring what we think we're measuring? We built QIMMA قمّة (Arabic for "summit"), to answer that question systematically. Instead of aggregating existing Arabic benchmarks as-is and running models on them, we applied a rigorous quality validation pipeline before any evaluation took place. What we found was sobering: even widely-used, well-regarded Arabic benchmarks contain systematic quality issues that can quietly corrupt evaluation results. This post walks through what QIMMA is, how we built it, what problems we found, and what the model rankings…
6dResearch#benchmark
34d ago
A New Framework for Evaluating Voice Agents (EVA)
A New Framework for Evaluating Voice Agents (EVA) Introduction Conversational voice agents present a distinct evaluation challenge: they must simultaneously satisfy two objectives — accuracy (completing the user's task correctly and faithfully) and conversational experience (doing so naturally, concisely, and in a way appropriate for spoken interaction). These objectives are deeply intertwined: mishearing a confirmation code renders perfect LLM reasoning meaningless, a wall of options overwhelms a caller who can't skim spoken output, and delayed responses can pass every accuracy check while remaining unusable in practice. Existing frameworks treat these as separate concerns — evaluating task success or conversational dynamics, but not both. We introduce EVA, an end-to-end evaluation framework for conversational voice agents that evaluates complete, multi-turn spoken conversations using a realistic bot-to-bot architecture. EVA produces two high-level scores, EVA-A (Accuracy) and EVA-X (Experience), and is designed to surface…
34dResearch#coding
38d ago
Build a Domain-Specific Embedding Model in Under a Day
Build a Domain-Specific Embedding Model in Under a Day With a single GPU and less than a day of training time, you can transform a general-purpose embedding model into one that truly understands your domain, no manual labeling required. To help you hit the ground running, we are also releasing a ready-to-use synthetic training dataset generated from NVIDIA's public documentation using this exact pipeline. Using this data and the recipe, we saw over 10% improvement in both Recall@10 and NDCG@10. Atlassian applied this recipe to fine-tune on their JIRA dataset, increasing Recall@60 from 0.751 to 0.951, a 26% improvement - on a single GPU. 🔗Quick Links to Dataset and Code: 🧑💻Open Source Projects Recipe Integrates: - NeMo Data Designer for synthetic data generation - NeMo Automodel for embedding model training - BEIR for Information retrieval evaluation - NeMo Export-Deploy for…
49d ago
Ulysses Sequence Parallelism: Training with Million-Token Contexts
Ulysses Sequence Parallelism: Training with Million-Token Contexts Ulysses Sequence Parallelism (part of the Arctic Long Sequence Training (ALST) protocol from Snowflake AI Research) provides an elegant solution by distributing the attention computation across multiple GPUs through attention head parallelism. In this post, we'll explore how Ulysses works and how it's been integrated across the Hugging Face ecosystem—from Accelerate to the Transformers Trainer and TRL's SFTTrainer. Contents - The Challenge of Long Sequence Training - How Ulysses Works - Integration with Accelerate - Integration with Transformers Trainer - Integration with TRL's SFTTrainer - Comparing Ulysses and Ring Attention - Best Practices - Benchmarks - Resources The Challenge of Long Sequence Training The attention mechanism in transformers scales quadratically with sequence length. For a sequence of length , standard attention requires FLOPs and memory to compute and store the attention score matrix.…
74d ago
OpenEnv in Practice: Evaluating Tool-Using Agents in Real-World Environments
OpenEnv in Practice: Evaluating Tool-Using Agents in Real-World Environments OpenEnv is an open-source framework from Meta and Hugging Face designed to address this challenge by standardizing how agents interact with real environments. As part of this collaboration, Turing contributed a production-grade calendar management environment to study tool-using agents under realistic constraints such as access control, temporal reasoning, and multi-agent coordination. In this post, we explore how OpenEnv works in practice, why calendars serve as a powerful benchmark for real-world agent evaluation, and what our findings reveal about the current limitations of tool-using agents. What Is OpenEnv? OpenEnv is a framework for evaluating AI agents against real systems rather than simulations. It provides a standardized way to connect agents to real tools and workflows while preserving the structure needed for consistent and reliable evaluation. OpenEnv uses a gym-oriented API (reset ,…
82d ago
Community Evals: Because we're done trusting black-box leaderboards over the community
Community Evals: Because we're done trusting black-box leaderboards over the community TL;DR: Benchmark datasets on Hugging Face can now host leaderboards. Models store their own eval scores. Everything links together. The community can submit results via PR. Verified badges prove that the results can be reproduced. Evaluation is broken Let's be real about where we are with evals in 2026. MMLU is saturated above 91%. GSM8K hit 94%+. HumanEval is conquered. Yet some models that ace benchmarks still can't reliably browse the web, write production code, or handle multi-step tasks without hallucinating, based on usage reports. There is a clear gap between benchmark scores and real-world performance. Furthermore, there is another gap within reported benchmark scores. Multiple sources report different results. From Model Cards, to papers, to evaluation platforms, there is no alignment in reported scores. The result is that…
82dResearch#benchmark
83d ago
H Company's new Holo2 model takes the lead in UI Localization
H Company's new Holo2 model takes the lead in UI Localization Two months since releasing our first batch of Holo2 models, H Company is back with our largest UI localization model yet: Holo2-235B-A22B Preview. This model achieves a new State-of-the-Art (SOTA) record of 78.5% on Screenspot-Pro and 79.0% on OSWorld G. Available on Hugging Face, Holo2-235B-A22B Preview is a research release focused on UI element localization. Agentic Localization High-resolution 4K interfaces are challenging for localization models. Small UI elements can be difficult to pinpoint on a large display. With agentic localization, however, Holo2 can iteratively refine its predictions, improving accuracy with each step and unlocking 10-20% relative gains across all Holo2 model sizes. Holo2-235B-A22B's Performance on ScreenSpot-Pro Holo2-235B-A22B Preview reaches 70.6% accuracy on ScreenSpot-Pro in a single step. In agent mode, it achieves 78.5% within 3 steps, setting a new…
83dResearch#agents
90d ago
Alyah ⭐️: Toward Robust Evaluation of Emirati Dialect Capabilities in Arabic LLMs
Alyah ⭐️: Toward Robust Evaluation of Emirati Dialect Capabilities in Arabic LLMs Arabic is one of the most widely spoken languages in the world, with hundreds of millions of speakers across more than twenty countries. Despite this global reach, Arabic is not a monolithic language. Modern Standard Arabic coexists with a rich landscape of regional dialects that differ significantly in vocabulary, syntax, phonology, and cultural grounding. These dialects are the primary medium of daily communication, oral storytelling, poetry, and social interaction. However, most existing benchmarks for Arabic large language models focus almost exclusively on Modern Standard Arabic, leaving dialectal Arabic largely under-evaluated and under-represented. This gap is particularly problematic as large language models increasingly interact with users in informal, culturally grounded, and conversational settings. A model that performs well on formal newswire text may still fail to understand a greeting,…
90dResearch
[IA(C]Import AI (Jack Clark)· 6 articlesvisit →
7d ago
Import AI 454: Automating alignment research; safety study of a Chinese model; HiFloat4
Import AI 454: Automating alignment research; safety study of a Chinese model; HiFloat4 At what point do the financial markets price in the singularity? Welcome to Import AI, a newsletter about AI research. Import AI runs on arXiv and feedback from readers. If you’d like to support this, please subscribe. Huawei’s HiFloat4 training format beats Western-developed MXFP4 in Ascend chip bakeoff: …Could this also be a symptom of the impact of export controls in driving Chinese interest towards maximizing training and inference efficiency? Perhaps… Huawei researchers have tested out HiFloat4, a 4-bit precision format for AI training and inference, against MXFP4, an Open Compute Project 4-bit format, and found that HiFloat4 is superior. This is interesting because it correlates to a broader level of interest in Chinese companies seeking to develop their own low-precision data formats explicitly coupled with their…
7dResearch#safetyby Jack Clark
14d ago
Import AI 453: Breaking AI agents; MirrorCode; and ten views on gradual disempowerment
Import AI 453: Breaking AI agents; MirrorCode; and ten views on gradual disempowerment Was fire equivalent to a singularity for people at the time? Welcome to Import AI, a newsletter about AI research. Import AI runs on arXiv and feedback from readers. If you’d like to support this, please subscribe. A shorter issue than usual as I was attending the 2026 Bilderberg conference this week. AI can reverse engineer software that contains thousands of lines of code: …MirrorCode demonstrates some of the long-horizon capabilities of modern AI systems… AI measurement organizations METR and Epoch have built MirrorCode, a benchmark meant to test out how well AI models can autonomously reimplement complex existing software. The results show that AI systems are more capable than most people think at certain types of coding task, suggesting AI progress may be even faster than…
14dResearch#agents#coding#benchmarkby Jack Clark
28d ago
Import AI 451: Political superintelligence; Google's society of minds, and a robot drummer
Import AI 451: Political superintelligence; Google's society of minds, and a robot drummer Are there any genies that can be put back in the bottle? Welcome to Import AI, a newsletter about AI research. Import AI runs on arXiv and feedback from readers. If you’d like to support this, please subscribe. AI might let us build “political superintelligence”: …But turning this into a societal upside requires lots of intentional work… As AI systems get more powerful and broaden their real world impact from coding to other domains, it seems likely that they could also become useful for helping people advocate for themselves in politics, and helping politicians better craft policy. But getting to a world where a “political superintelligence” exists and helps us is a lot more challenging than just building better AI systems, according to Andy Hall, a political…
28dResearch#codingby Jack Clark
63d ago
Import AI 446: Nuclear LLMs; China's big AI benchmark; measurement and AI policy
Import AI 446: Nuclear LLMs; China's big AI benchmark; measurement and AI policy Will AIs be jealous of one another? Welcome to Import AI, a newsletter about AI research. Import AI runs on arXiv and feedback from readers. If you’d like to support this, please subscribe. Want to make AI go better? Figure out how to measure it: …One simple policy intervention that works well… Jacob Steinhardt, an AI researcher, has written a nice blog laying out the virtues in investing in technical tools to measure properties of AI systems and drive down costs in complying with technical policy solutions. As someone who has spent their professional life in AI writing about AI measurement and building teams (e.g, the Frontier Red Team and Societal Impacts and Economic Research teams at Anthropic) to measure properties of AI systems, I agree with…
63dResearch#benchmarkby Jack Clark
70d ago
Import AI 445: Timing superintelligence; AIs solve frontier math proofs; a new ML research benchmark
Import AI 445: Timing superintelligence; AIs solve frontier math proofs; a new ML research benchmark Will 2026 be looked back on as the pivotal year for making decisions about the singularity? Welcome to Import AI, a newsletter about AI research. Import AI runs on arXiv and feedback from readers. If you’d like to support this, please subscribe. Economist: Don’t worry about AI-driven unemployment, because people like paying for the ‘human touch’: …Even when you have the technology to automate something, you might still pick a human…Adam Ozimek, chief economist at the Economic Innovation Group, has written a blog noting that even if AI gets much, much better and is capable of doing all the work that people do, there will still be some jobs for humans because people seem to have a preference for humans over machines in certain domains.…
70dResearch#benchmarkby Jack Clark
77d ago
Import AI 444: LLM societies; Huawei makes kernels with AI; ChipBench
Import AI 444: LLM societies; Huawei makes kernels with AI; ChipBench How can you quantify creativity? Welcome to Import AI, a newsletter about AI research. Import AI runs on arXiv and feedback from readers. If you’d like to support this, please subscribe. Google paper suggests that LLMs simulate multiple personalities to answer questions: …The smarter we make language models, the more they tend towards building and manipulating rich, multi-agent world models… When thinking about hard problems, I often find it’s helpful to try and view them from multiple perspectives, especially when it comes to checking my own assumptions and biases. Now, researchers with Google, the University of Chicago, and the Santa Fe Institute, have studied how AI reasoning models work and have concluded they do the same thing, with LLMs seeming to invoke multiple different perspectives in their chains of…
77dResearch#agents#safetyby Jack Clark
[MRB]Microsoft Research Blog· 8 articlesvisit →
7d ago
Can we AI our way to a more sustainable world?
Technical advancement is moving at such a rapid pace that it can be challenging to define the tomorrow we’re working toward. In The Shape of Things to Come, Microsoft Research leader Doug Burger and experts from across disciplines tease out the thorniest AI issues facing technologists, policymakers, business decisionmakers, and other stakeholders today. The goal: to amplify the shared understanding needed to build a future in which the AI transition is a net positive. In this episode, Burger is joined by Amy Luers, head of sustainability science and innovation at Microsoft, and Ishai Menache, an optimization researcher at Microsoft Research, to explore how AI can both contribute to and help address climate change, emphasizing the need to separate hype from data and understand its real impact. While datacenters account for a small share of global emissions, their rapid growth raises…
7dResearchby Doug Burger, Amy Luers, Ishai Menache
18d ago
Ideas: Steering AI toward the work future we want
Behind every emerging technology is a great idea propelling it forward. In the Microsoft Research Podcast series Ideas, members of the research community at Microsoft discuss the beliefs that animate their research, the experiences and thinkers that inform it, and the positive human impact it targets. Since 2020, researchers across Microsoft have conducted, surfaced, and analyzed key research into how people work as part of the New Future of Work research initiative. They’ve done this through a variety of lenses—from changes caused by the pandemic to the adoption of hybrid work practices to the arrival of increasingly capable AI models—with the goal of empowering people and organizations to redefine work in real time. In this episode, Microsoft Chief Scientist and Technical Fellow Jaime Teevan talks with researchers Jenna Butler, Jake Hofman, and Rebecca Janssen about the latest efforts: the Microsoft…
18dResearchby Jaime Teevan, Jenna Butler, Jake Hofman, Rebecca Janssen
18d ago
New Future of Work: AI is driving rapid change, uneven benefits
At a glance - AI is driving rapid changes in the workplace, more sharply than those covered in previous editions of the New Future of Work - AI is changing how people work together, not just enabling them to work faster or from remote locations. Organizations that treat AI as a collaborative partner are seeing the biggest benefits. - The benefits of AI are not yet evenly distributed, underscoring the need for industry leaders to build AI that expands opportunity. The future is not predetermined. It will be shaped by the choices we make today. - Human expertise matters more, not less, in an AI-powered world. People are shifting from merely doing work to guiding, critiquing, and improving the work of AI. For the past five years, the New Future of Work report has captured how work is changing. This…
18dResearchby Jaime Teevan, Sonia Jaffe, Rebecca Janssen, Nancy Baym, Siân Lindley, Bahar Sarrafzadeh, Brent Hecht, Jenna Butler, Jake Hofman, Sean Rintel
26d ago
ADeLe: Predicting and explaining AI performance across tasks
At a glance - AI benchmarks report performance on specific tasks but provide limited insight into underlying capabilities; ADeLe evaluates models by scoring both tasks and models across 18 core abilities, enabling direct comparison between task demands and model capabilities. - Using these ability scores, the method predicts performance on new tasks with ~88% accuracy, including for models such as GPT-4o and Llama-3.1. - It builds ability profiles and identifies where models are likely to succeed or fail, highlighting strengths and limitations across tasks. - By linking outcomes to task demands, ADeLe explains differences in performance, showing how it changes as task complexity increases. AI benchmarks report how large language models (LLMs) perform on specific tasks but provide little insight into their underlying capabilities that drive their performance. They do not explain failures or reliably predict outcomes on new tasks.…
26dResearch#benchmarkby Lexin Zhou, Xing Xie
32d ago
GroundedPlanBench: Spatially grounded long-horizon task planning for robot manipulation
At a glance - VLM-based robot planners struggle with long, complex tasks because natural-language plans can be ambiguous, especially when specifying both actions and locations. - GroundedPlanBench evaluates whether models can plan actions and determine where they should occur across diverse, real-world robot scenarios. - Video-to-Spatially Grounded Planning (V2GP) is a framework that converts robot demonstration videos into spatially grounded training data, enabling models to learn planning and grounding jointly. - Grounded planning improves both task success and action accuracy, outperforming decoupled approaches in benchmark and real-world evaluations. Vision-language models (VLMs) use images and text to plan robot actions, but they still struggle to decide what actions to take and where to take them. Most systems split these decisions into two steps: a VLM generates a plan in natural language, and a separate model translates it into executable actions. This…
32dResearch#multimodalby Sehun Jung, HyunJee Song, Dong-Hee Kim, Reuben Tan, Jianfeng Gao, Yong Jae Lee, Donghyun Kim
32d ago
AsgardBench: A benchmark for visually grounded interactive planning
At a glance - To successfully complete tasks, embodied AI agents must ground and update their plans based on visual feedback. - AsgardBench isolates whether agents can use visual observations to revise their plans as tasks unfold. - Spanning 108 controlled task instances across 12 task types, the benchmark requires agents to adapt their plans based on what they observe. - Because objects can be in different positions and states (e.g., clean or dirty), the same instruction can require different action sequences, even in the same environment. Imagine a robot tasked with cleaning a kitchen. It needs to observe its environment, decide what to do, and adjust when things don’t go as expected, for example, when the mug it was tasked to wash is already clean, or the sink is full of other items. This is the domain of embodied…
32dResearch#benchmarkby Andrea Tupini, Lars Liden, Reuben Tan, Yu Wang, Jianfeng Gao
35d ago
Will machines ever be intelligent?
Technical advances are moving at such a rapid pace that it can be challenging to define the tomorrow we’re working toward. In The Shape of Things to Come, Microsoft Research leader Doug Burger and experts from across disciplines tease out the thorniest AI issues facing technologists, policymakers, business decision-makers, and other stakeholders today. The goal: to amplify the shared understanding needed to build a future in which the AI transition is a net positive. In this first episode of the series, Burger is joined by Nicolò Fusi of Microsoft Research and Subutai Ahmad (opens in new tab) of Numenta to examine whether today’s AI systems are truly intelligent. They compare transformer-based large language models (LLMs) with the human brain’s distributed, continuously learning architecture, exploring differences in efficiency, representation, and sensory-motor grounding. The discussion probes what intelligence really means, where current…
35dResearchby Doug Burger, Subutai Ahmad, Nicolo Fusi
46d ago
Systematic debugging for AI agents: Introducing the AgentRx framework
At a glance - Problem: Debugging AI agent failures is hard because trajectories are long, stochastic, and often multi-agent, so the true root cause gets buried. - Solution: AgentRx (opens in new tab) pinpoints the first unrecoverable (“critical failure”) step by synthesizing guarded, executable constraints from tool schemas and domain policies, then logging evidence-backed violations step-by-step. - Benchmark + taxonomy: We release AgentRx Benchmark (opens in new tab) with 115 manually annotated failed trajectories across τ-bench, Flash, and Magentic-One, plus a grounded nine-category failure taxonomy. - Results + release: AgentRx improves failure localization (+23.6%) and root-cause attribution (+22.9%) over prompting baselines, and we are open-sourcing the framework and dataset. As AI agents transition from simple chatbots to autonomous systems capable of managing cloud incidents, navigating complex web interfaces, and executing multi-step API workflows, a new challenge has emerged: transparency. When…
46dResearch#agentsby Shraddha Barke, Arnav Goyal, Alind Khare, Chetan Bansal
[MTR]MIT Technology Review· 3 articlesvisit →
3d ago
The Download: supercharged scams and studying AI healthcare
The Download: supercharged scams and studying AI healthcare Plus: DeepSeek has unveiled its long-awaited new AI model. This is today's edition of The Download, our weekday newsletter that provides a daily dose of what's going on in the world of technology. We’re in a new era of AI-driven scams When ChatGPT was released in late 2022, it showed how easily generative AI could create human-like text. This quickly caught the eye of cybercriminals, who began using LLMs to compose malicious emails. Since then, they’ve adopted AI for everything from turbocharged phishing and hyperrealistic deepfakes to automated vulnerability scans. Many organizations are now struggling to cope with the sheer volume of cyberattacks. AI is making them faster, cheaper, and easier to carry out, a problem set to worsen as more cybercriminals adopt these tools—and their capabilities improve. Read the full story…
3dResearch#gptby Thomas Macaulay
4d ago
Will fusion power get cheap? Don’t count on it.
Will fusion power get cheap? Don’t count on it. New research suggests that cost declines could be slow for the technology. Fusion power could provide a steady, zero-emissions source of electricity in the future—if companies can get plants built and running. But a new study suggests that even if that future arrives, it might not come cheap. Technologies tend to get less expensive over time. Lithium-ion batteries are now about 90% cheaper than they were in 2013. But historically, different technologies tend to go through this curve at different rates. And the cost of fusion might not sink as quickly as the prices of batteries or solar. It’s tricky to make any predictions about the cost of a technology that doesn’t exist yet. But when there’s billions of dollars of public and private funding on the line, it’s worth considering…
4dResearchby Casey Crownhart
5d ago
AI needs a strong data fabric to deliver business value
Sponsored AI needs a strong data fabric to deliver business value A modern data fabric makes it possible to turn existing enterprise knowledge into a trusted foundation for AI. In partnership withSAP Artificial intelligence is moving quickly in the enterprise, from experimentation to everyday use. Organizations are deploying copilots, agents, and predictive systems across finance, supply chains, human resources, and customer operations. By the end of 2025, half of companies used AI in at least three business functions, according to a recent survey. But as AI becomes embedded in core workflows, business leaders are discovering that the biggest obstacle is not model performance or computing power but the quality and the context of the data on which those systems rely. AI essentially introduces a new requirement: Systems must not only access data — they must understand the business context behind…
5dResearch#codingby MIT Technology Review Insights
[NL(]Nathan Lambert (RLHF)· 3 articlesvisit →
8d ago
Announcing RAAIS 2026 headline speakers
Announcing RAAIS 2026 headline speakers Frontier AI, open-ended agents, AI for medicine, world models, and the first data centers in orbit, all at the 10th Research and Applied AI Summit on June 12th, 2026. The Research and Applied AI Summit (RAAIS) is a community for entrepreneurs and researchers who accelerate the science and applications of AI technology. The 10th annual summit takes place on June 12th, 2026 in London. We’re delighted to announce the first wave of headline speakers, across five threads: frontier AI, open-ended agents, AI for medicine and science, world models, and the next substrate for compute itself. Frontier AI and the future of intelligence Raia Hadsell is VP of Research at Google DeepMind, where she co-leads the Frontier AI unit and has contributed to Gemini 2.5, Gemma 2, RecurrentGemma, and RoboCat. Her earlier seminal work includes Overcoming…
8dResearchby Nathan Benaich
15d ago
State of AI: April 2026 newsletter
State of AI: April 2026 newsletter US Government blacklists Anthropic as Iran bombs AWS data centers. Plus: $19B revenue in weeks, industrial-scale distillation wars, and an mRNA dog cancer vaccine designed by ChatGPT. Dear readers, Welcome to the latest issue of the State of AI, an editorialized newsletter that covers the key developments in AI policy, research, industry, and start-ups from February 1 to April 7, 2026. First up, a few news items: Air Street Capital Epoch 3 is live! $232M to continue backing AI-first companies across the US and Europe in software, dev/infra, techbio and defense. RAAIS 2026 is back in London on June 12. This year’s speakers include Raia Hadsell (VP Research, Google DeepMind), Roberta Raileanu (Senior Staff Research Scientist, Google DeepMind), Jeff Hawke (Co-Founder & CTO, Odyssey), and Philip Johnston (Co-Founder & CEO, Starcloud - yes, data…
15dResearch#gptby Nathan Benaich
35d ago
Air Street Capital announces $232M Fund III to back AI-first companies
Air Street Capital announces $232M Fund III to back AI-first companies Entering our third epoch. Today, I’m thrilled to share that Air Street Capital has raised a third fund of $232,323,232 to back AI-first companies from the earliest stages in North America and Europe. Air Street will lead early stage rounds with checks of $500k to $15M and make select growth investments up to $25M. When I started investing in 2013, deep learning was largely confined to research labs. Yet I was convinced back then that the most important technology companies of our generation will be built AI-first. This is because AI is a force multiplier for technological progress, and everything around us is ultimately a product of intelligence. So in 2019, I founded Air Street to build a venture firm dedicated to this conviction. Today, AI-first companies are emerging…
35dResearchby Nathan Benaich
[NV]NVIDIA Developer Blog· 8 articlesvisit →
3d ago
Federated Learning Without the Refactoring Overhead Using NVIDIA FLARE
Federated learning (FL) is no longer a research curiosity—it’s a practical response to a hard constraint: the most valuable data is often the least movable. Regulatory boundaries, data sovereignty rules, and organizational risk tolerance routinely prevent centralized aggregation. Meanwhile, sheer data gravity makes even permitted transfers slow, expensive, and fragile at scale. The latest version of NVIDIA FLARE addresses this reality with a federated computing runtime that moves the training logic to the data, while raw data stays put. In high-stakes environments, centrally aggregating data is often not possible or practical, so a modern federated platform must treat data isolation, compliance, and privacy-enhancing technologies as first-class requirements. What has historically slowed adoption isn’t the concept of FL—it’s the developer experience. If the path from “my local script trains” to “my job runs across federated sites” requires deep refactoring, new class…
3dResearch#gpuby Holger Roth
4d ago
Winning a Kaggle Competition with Generative AI–Assisted Coding
In March 2026, three LLM agents generated over 600,000 lines of code, ran 850 experiments, and helped secure a first-place finish in a Kaggle playground competition. Success in modern machine learning competitions is increasingly defined by how quickly you can generate, test, and iterate on ideas. LLM agents, combined with GPU acceleration, dramatically compress this loop. Historically, two bottlenecks have limited this experimentation: - How quickly you can write code for new experiments. - How quickly you can execute those experiments. GPUs and libraries like NVIDIA cuDF, NVIDIA cuML, XGBoost, and PyTorch have largely solved the second problem. LLM agents now address the first problem—unlocking a new scale of rapid, iterative experimentation. This blog post describes how I used LLM agents to accelerate the discovery of the most performant tabular data prediction solutions. Case study: Kaggle Playground churn prediction The…
4dResearch#codingby Chris Deotte
10d ago
Accelerate Clean, Modular, Nuclear Reactor Design with AI Physics
The development of socially acceptable nuclear reactors requires that they are safe, clean, efficient, economical, and sustainable. Meeting these requirements calls for new approaches, driving growing interest in Small Modular Reactors (SMRs) and in Generation IV designs. SMRs aim to improve project economics by standardising designs and shifting construction to controlled manufacturing environments, while Gen IV reactors target fundamental fuel-cycle challenges by better managing transuranics and reducing the radiotoxicity and longevity of waste. Together, these approaches offer a credible roadmap toward safer, cleaner, and more sustainable nuclear energy. However, validating new designs presents significant challenges. Due to the expense, time constraints, and inherent complexities of physical experiments, numerical simulations are fundamental to the design of nuclear reactors. Yet, the high computational cost of these simulations often creates a major bottleneck in the design process, slowing the pace of innovation. To…
10dResearchby Mark Hobbs
13d ago
Building Custom Atomistic Simulation Workflows for Chemistry and Materials Science with NVIDIA ALCHEMI Toolkit
For decades, computational chemistry has faced a tug-of-war between accuracy and speed. Ab initio methods like density functional theory (DFT) provide high fidelity but are computationally expensive, limiting researchers to systems of a few hundred atoms. Conversely, classical force fields are fast but often lack the chemical accuracy required for complex bond-breaking or transition-state analysis. Machine learning interatomic potentials (MLIPs) have emerged as the bridge, offering quantum accuracy at classical speeds. However, the software ecosystem is a new bottleneck. While the MLIP models themselves run on GPUs, the surrounding simulation infrastructure often relies on legacy CPU-centric code. NVIDIA ALCHEMI (AI Lab for Chemistry and Materials Innovation) helps to address these challenges by accelerating chemicals and materials discovery with AI. We have previously announced two components of the ALCHEMI portfolio: - ALCHEMI NIM microservices: Scalable, cloud‑ready microservices for AI-accelerated batched atomistic…
13dResearch#agents#gpuby Erica Tsai
35d ago
Building a Zero-Trust Architecture for Confidential AI Factories
AI is moving from experimentation to production. However, most data enterprises need exists outside the public cloud. This includes sensitive information like patient records, market research, and legacy systems containing enterprise knowledge. There’s also a risk of using private data with AI models, and adoption is often slowed or blocked by privacy and trust concerns. Enterprises building next-generation AI factories—specializing in high-performance infrastructure to manufacture intelligence at scale—must be built on a zero-trust foundation. This security architecture eliminates implicit trust in the underlying host infrastructure by using hardware-enforced Trusted Execution Environments (TEEs) and cryptographic attestation. This post describes the full-stack architecture needed to integrate the zero-trust foundation into AI factories. On-premise requirements often limit enterprises to building their own models or using open source models for agentic AI workloads. To deliver on the promise of AI, organizations must deploy a…
35dResearchby Hema Bontha
68d ago
Topping the GPU MODE Kernel Leaderboard with NVIDIA cuda.compute
Python dominates machine learning for its ergonomics, but writing truly fast GPU code has historically meant dropping into C++ to write custom kernels and to maintain bindings back to Python. For most Python developers and researchers, this is a significant barrier to entry. Frameworks like PyTorch address this by implementing kernels in CUDA C++—either handwritten or by leveraging libraries like the NVIDIA CUDA Core Compute Libraries. Handwritten kernels are time-consuming and require deep, low-level architectural expertise. Using CUB, a C++ library within CCCL, is often better, since its primitives are highly optimized per architecture and are rigorously tested. But exposing CUB to Python traditionally means building and maintaining bindings and pre-instantiating C++ templates with fixed types and operators—limiting flexibility on the Python side. The NVIDIA cuda.compute library overcomes these limitations by offering a high-level, Pythonic API for device-wide CUB primitives.…
68dResearch#coding#benchmark#gpuby Daniel Rodriguez
76d ago
Using Accelerated Computing to Live-Steer Scientific Experiments at Massive Research Facilities
Scientists and engineers who design and build unique scientific research facilities face similar challenges. These include managing massive data rates that exceed current computational infrastructure capacity to extract scientific insights and driving the experiments in real time. These challenges are obstacles to maximizing the impact of scientific discoveries and significantly slow the pace of knowledge growth. Scientists and engineers at NVIDIA work with these facilities to develop new solutions built on parallel and distributed computation that remove these blockers. This post will walk through two notable examples of formalizing complex physics problems into tractable mathematical puzzles that benefit greatly from GPU-accelerated scientific computing, involving the U.S. Department of Energy: NSF-DOE Vera C. Rubin Observatory and SLAC’s Linac Coherent Light Source II (LCLS-II). These unique and massive-scale research facilities both took a decade to build and enable unprecedented scientific discoveries to…
76dResearchby Quynh L. Nguyen
87d ago
Establishing a Scalable Sparse Ecosystem with the Universal Sparse Tensor
Sparse tensors are vectors, matrices, and higher-dimensional generalizations with many zeros. They are crucial in various fields such as scientific computing, signal processing, and deep learning due to their efficiency in storage, computation, and power. Despite their benefits, handling sparse tensors manually or through existing libraries is often cumbersome, error-prone, nonportable, and does not scale with the combinatorial explosion of sparsity patterns, data types, operations, and targets. Research largely focuses on sparse storage formats—data structures that compactly store nonzeros and allow efficient operations that avoid redundancies such as x+0=x and x*0=0. This enables scaling to larger sizes or solving same sizes with fewer resources. No single sparse format is optimal; the best choice depends on the nonzero distribution, operations, and target architecture. The Universal Sparse Tensor (UST) decouples a tensor’s sparsity from its memory storage representation. The UST uses a…
87dResearch#rag#embeddingsby Aart J.C. Bik
[OAI]OpenAI Blog· 34 articlesvisit →
4d ago
Introducing GPT-5.5
Update on April 24, 2026: GPT‑5.5 and GPT‑5.5 Pro are now available in the API. The system card has also been updated to describe the additional safeguards that apply. We’re releasing GPT‑5.5, our smartest and most intuitive to use model yet, and the next step toward a new way of getting work done on a computer. GPT‑5.5 understands what you’re trying to do faster and can carry more of the work itself. It excels at writing and debugging code, researching online, analyzing data, creating documents and spreadsheets, operating software, and moving across tools until a task is finished. Instead of carefully managing every step, you can give GPT‑5.5 a messy, multi-part task and trust it to plan, use tools, check its work, navigate through ambiguity, and keep going. The gains are especially strong in agentic coding, computer use, knowledge work,…
4dResearch#coding
5d ago
Making ChatGPT better for clinicians
Making ChatGPT better for clinicians Built for clinical work, ChatGPT for Clinicians is now available for free to verified individual clinicians in the U.S. We’re introducing ChatGPT for Clinicians, a version of ChatGPT designed to support clinical tasks like documentation and medical research so clinicians can focus on delivering high-quality patient care. We’re making it free for any verified physician, NP, PA, or pharmacist, starting in the U.S. The U.S. healthcare system today is under extraordinary strain. Clinicians are being asked to care for more patients while managing growing administrative demands and a rapidly expanding body of medical research. Many are already turning to AI tools like ChatGPT for support. According to a 2026 survey by the American Medical Association(opens in a new window), physician use of AI is now at an all-time high, with 72% of physicians reporting they…
5dResearch#gpt
11d ago
Introducing GPT-Rosalind for life sciences research
Introducing GPT‑Rosalind for life sciences research A new purpose-built model to accelerate scientific research and drug discovery. Today, we’re introducing GPT‑Rosalind, our frontier reasoning model built to support research across biology, drug discovery, and translational medicine. The life sciences model series is optimized for scientific workflows, combining improved tool use with deeper understanding across chemistry, protein engineering, and genomics. On average, it takes roughly 10 to 15 years to go from target discovery to regulatory approval for a new drug in the United States. Gains made at the earliest stages of discovery compound downstream in better target selection, stronger biological hypotheses and higher-quality experiments. Progress in the life sciences is constrained not only by the difficulty of the underlying science, but by the complexity of the research workflows themselves. Scientists must work across large volumes of literature, specialized databases, experimental…
11dResearch#agents
17d ago
Applications of AI at OpenAI
Applications of AI at OpenAI Explore how OpenAI products and APIs bring AI into real-world use. OpenAI was founded with a long-term goal: to ensure advanced AI benefits humanity. Early work focused on research and experimentation, followed by large-scale model development. Over time, OpenAI began releasing models through both consumer-facing products and developer platforms, allowing individuals, teams, and organizations to apply AI to their work. At a high level, OpenAI currently supports AI applications in two ways: 1) Direct access through OpenAI products, like ChatGPT or Codex. These are tools people can use immediately for learning, work, creativity, and building. 2) Composable building blocks through APIs. These allow developers to integrate model intelligence into their own workflows, products, and systems. The sections below summarize the most common OpenAI products and what they’re designed for. ChatGPT is OpenAI’s main user-facing product—a…
21d ago
Industrial policy for the Intelligence Age
Industrial policy for the Intelligence Age Ideas to keep people first. As we move toward superintelligence, incremental policy updates won’t be enough. To kick-start this much needed conversation, OpenAI is offering a slate of people-first policy ideas(opens in a new window) designed to expand opportunity, share prosperity, and build resilient institutions—ensuring that advanced AI benefits everyone. These ideas are ambitious, but intentionally early and exploratory. We offer them not as a comprehensive or final set of recommendations, but as a starting point for discussion that we invite others to build on, refine, challenge, or choose among through the democratic process. To help sustain momentum, OpenAI is: - welcoming and organizing feedback through newindustrialpolicy@openai.com - establishing a pilot program of fellowships and focused research grants of up to $100,000 and up to $1 million in API credits for work that builds…
21dResearch#fine-tuning
21d ago
Announcing the OpenAI Safety Fellowship
Introducing the OpenAI Safety Fellowship A pilot program to support independent safety and alignment research and develop the next generation of talent Today we are announcing a call for applications to the OpenAI Safety Fellowship, a new program for external researchers, engineers, and practitioners to pursue rigorous, high-impact research on the safety and alignment of advanced AI systems. The program will run from September 14, 2026 through February 5, 2027. We are looking for applicants interested in safety questions that matter for existing and future systems. Priority areas include safety evaluation, ethics, robustness, scalable mitigations, privacy-preserving safety methods, agentic oversight, and high-severity misuse domains, among others. We are especially interested in work that is empirically grounded, technically strong, and relevant to the broader research community. Fellows will work closely with OpenAI mentors and engage with a cohort of peers. Workspace…
21dResearch#safety
39d ago
How we monitor internal coding agents for misalignment
How we monitor internal coding agents for misalignment Using our most powerful models to detect and study misaligned behavior in real-world deployments. AI systems are beginning to act with greater autonomy in real-world environments at scale. As their capabilities advance, they are able to take on increasingly complex, high-impact tasks and interact with tools, systems, and workflows in ways that resemble human collaborators. A core part of OpenAI’s mission is helping the world navigate this transition to AGI responsibly. That means not only building highly capable systems, but also developing the methods, infrastructure, and approaches needed to deploy and manage them safely as their capabilities continue to grow. Monitoring internally deployed agents is one of the key ways we’re doing this, and it allows us both to learn from real-world usage and to identify and mitigate emerging risks. Over the…
41d ago
Equipping workers with insights about compensation
Equipping workers with insights about compensation Americans are sending nearly 3 million messages to ChatGPT each day to help them close the wage information gap. Wage information shapes important decisions: what jobs people apply for, whether they negotiate, and whether a particular career path is worth pursuing. But unlike the price of most goods, the price of labor is often hard to find and difficult to interpret—especially for workers who are early in their careers, switching fields, or moving locations. AI is a new type of labor-market resource. Rather than requiring a worker to search across multiple websites, interpret scattered salary pages, or ask a socially risky question, a model can synthesize wage information and return a benchmark in seconds. Workers are already using ChatGPT this way, sending nearly 3 million messages per day, on average in the US, asking…
41dResearch#gpt
42d ago
Why Codex Security Doesn’t Include a SAST Report
For decades, static application security testing (SAST) has been one of the most effective ways security teams scale code review. But when we built Codex Security, we made a deliberate design choice: we didn’t start by importing a static analysis report and asking the agent to triage it. We designed the system to start with the repository itself—its architecture, trust boundaries, and intended behavior—and to validate what it finds before it asks a human to spend time on it. The reason is simple: the hardest vulnerabilities usually aren’t dataflow problems. They happen when code appears to enforce a security check, but that check doesn’t actually guarantee the property the system relies on. In other words, the challenge isn’t just tracking how data moves through a program—it’s determining whether the defenses in the code really work. SAST is often framed as…
42dResearch#agents#coding
47d ago
Wayfair boosts catalog accuracy and support speed with OpenAI
Wayfair boosts catalog accuracy and support speed with OpenAI By embedding OpenAI models in supplier and catalog systems, Wayfair improved data accuracy and automated workflows for millions of products. Results 2.5M Product tags corrected Results 41K Supplier support tickets automated per month Results 1,200 ChatGPT Enterprise seats deployed Wayfair, one of the world’s largest home goods retailers, has integrated OpenAI models into critical internal systems to improve supplier support workflows and product catalog quality at scale. What began as value-testing small scale releases in 2024 has evolved into a full production system that reduces manual effort, accelerates decision-making and improves data quality across millions of products. Rather than treat generative AI as an experiment or point solution, Wayfair embedded OpenAI models into core operational workflows. The company focused first where complexity and need for scale were highest: routing and resolving…
49d ago
OpenAI to acquire Promptfoo
OpenAI to acquire Promptfoo Accelerating agentic security testing and evaluation capabilities in OpenAI Frontier We’re acquiring Promptfoo, an AI security platform that helps enterprises identify and remediate vulnerabilities in AI systems during development. Once the acquisition is finalized we will integrate Promptfoo’s technology directly into OpenAI Frontier, our platform for building and operating AI coworkers. As enterprises deploy AI coworkers into real workflows, evaluation, security, and compliance become foundational requirements. Enterprises need systematic ways to test agent behavior, detect risks before deployment, and maintain clear records to support oversight, governance, and accountability over time. The Promptfoo team, led by Ian Webster and Michael D’Angelo, has built a powerful suite of tools trusted by over 25 percent of Fortune 500 companies, along with a widely used open-source(opens in a new window) CLI and library for evaluating and red-teaming LLM applications. Together,…
52d ago
How Balyasny Asset Management built an AI research engine
How Balyasny Asset Management built an AI research engine By combining rigorous model evaluation, full-platform use of OpenAI, and agent workflows, Balyasny is reinventing investment research. Results 95% Portion of investment team using the AI research system Results Days to hours With agents powered by OpenAI models, deep research tasks that once required days are now completed in hours Balyasny Asset Management(opens in a new window) (Balyasny) is a global, multi-strategy investment firm with approximately 180 investment teams across diverse asset classes and geographies. The firm operates in a highly competitive and dynamic industry where conviction, precision, and speed are all critical to success. Facing an increasingly complex market environment with surging volumes of financial data, Balyasny saw an opportunity to reimagine the investment research process using AI. In late 2022, Balyasny established an Applied AI team: a centralized group…
52dResearch#agents
52d ago
Codex Security: now in research preview
Today we’re introducing Codex Security, our application security agent. It builds deep context about your project to identify complex vulnerabilities that other agentic tools miss, surfacing higher-confidence findings with fixes that meaningfully improve the security of your system while sparing you from the noise of insignificant bugs. Context is essential when evaluating real security risks, but most AI security tools simply flag low-impact findings and false positives, forcing security teams to spend significant time on triage. At the same time, agents are accelerating software development, making security review an increasingly critical bottleneck. Codex Security addresses both challenges. By combining agentic reasoning from our frontier models with automated validation, it delivers high-confidence findings and actionable fixes so teams can focus on the vulnerabilities that matter and ship secure code faster. Formerly known as Aardvark, Codex Security began last year as a…
52dResearch#agents
53d ago
The five AI value models driving business reinvention
The five AI value models driving business reinvention Most organizations still manage AI as a series of use cases: a pilot here, a workflow there, a promising tool inside one function. That approach can generate local wins but it rarely transforms how a business creates value. It is akin to creating interactive banners and drip email campaigns with the arrival of the internet, and missing the point of the eCommerce revolution. The organizations pulling ahead use a different, and more ambitious logic. They treat AI not as a collection of disconnected experiments, but as a portfolio of value models. Each has its own economics, time-to-value, and governance requirements, and each makes the next one easier to scale. This is why the companies that get the most from AI will not be the ones running the most pilots. They will be…
53dResearch#agents#local
53d ago
Introducing ChatGPT for Excel and new financial data integrations
Introducing ChatGPT for Excel and new financial data integrations Use ChatGPT in Excel to build, update, and analyze spreadsheets faster, and new integrations in ChatGPT for financial workflows. Update on April 22, 2026: ChatGPT for Google Sheets is now available in beta, bringing ChatGPT into Google Sheets so users can build, analyze, and update spreadsheets using natural language. We've also added support for app integrations and skills for both ChatGPT for Excel and ChatGPT for Google Sheets. Learn more(opens in a new window). Today, we’re introducing ChatGPT for Excel(opens in a new window) in beta, an Excel add-in that brings ChatGPT directly into workbooks to help build and update models, run scenarios, and generate outputs based on cells and formulas. Powered by GPT‑5.4, it helps users do more in Excel, supports power users in moving faster, and can improve consistency…
53dResearch#gpt
53d ago
Reasoning models struggle to control their chains of thought, and that’s good
Reasoning models struggle to control their chains of thought, and that’s good Why a limitation of frontier models is reassuring for AI safety. As AI agents become capable of carrying out increasingly complex and autonomous tasks, maintaining reliable oversight of their behavior becomes more important. Consistent with our principle of iterative deployment, we study how systems behave in real-world settings and continuously refine safeguards as capabilities advance. To support this, our safety approach uses defense-in-depth, with multiple complementary layers of defense such as safety training, behavioral testing, agentic code review(opens in a new window), and chain-of-thought (CoT) monitoring. CoT monitoring analyzes the reasoning steps agents generate while pursuing tasks. These reasoning traces can provide valuable signals during both training and deployment, helping monitoring systems identify when an agent’s behavior may be unsafe or inconsistent with the user’s intended goals. Today,…
54d ago
How Axios uses AI to help deliver high-impact local journalism
How Axios uses AI to help deliver high-impact local journalism A conversation with Allison Murphy, Chief Operating Officer, Axios. Axios is a media company delivering vital, trustworthy news and analysis in the most efficient, illuminating and shareable ways possible. It offers a mix of original and smartly narrated coverage of media trends, tech, business and politics with expertise, voice and smart brevity. We spoke with Allison Murphy, Chief Operating Officer at Axios, about AI supporting high-impact local journalism and serving communities better. AI is already a huge part of how Axios Local works. At the core, what we’re trying to do is prove that you can run a sustainable, profitable local news model that delivers high-quality journalism to every community in America. That means solving for scale and efficiency—and that’s exactly what AI is good at. So there’s a really…
59d ago
Joint Statement from OpenAI and Microsoft
Joint Statement from OpenAI and Microsoft Since 2019, Microsoft and OpenAI have worked together to advance artificial intelligence responsibly and make its benefits broadly accessible. What began as a research partnership has grown into one of the most consequential collaborations in technology—grounded in mutual trust, deep technical integration, and a long‑term commitment to innovation. As conversations around AI investments and partnerships grow and as OpenAI announces new funding and new partners as they did today, we want to ensure these announcements are understood within the existing construct of our partnership. Nothing about today’s announcements in any way changes the terms of the Microsoft and OpenAI relationship that have been previously shared in our joint blog in October 2025. The partnership remains strong and central. Microsoft and OpenAI continue to work closely across research, engineering, and product development, building on years…
59dResearch
60d ago
Pacific Northwest National Laboratory and OpenAI partner to accelerate federal permitting
Pacific Northwest National Laboratory and OpenAI partner to accelerate federal permitting New benchmark shows potential to reduce infrastructure permitting timelines Modernizing how the federal government permits critical infrastructure is essential to building a faster, safer, and more competitive U.S. economy. From energy projects and advanced manufacturing to transportation and water systems, permitting determines how quickly promising ideas become real-world investments. Yet today, environmental and technical reviews often take years, which slows innovation, increases costs, and delays the benefits these projects deliver to communities. That’s why OpenAI has partnered with the U.S. Department of Energy’s Pacific Northwest National Laboratory (PNNL) and its PermitAI™(opens in a new window) team to evaluate whether coding agents can help effectively accelerate federal permitting work. PermitAI, an initiative funded by the Department of Energy’s Office of Policy, and OpenAI worked together with 19 subject matter experts…
61d ago
Personalizing education with ChatGPT
Arizona State University personalizes learning and advances research with ChatGPT Arizona State University(opens in a new window) (ASU) is one of the largest public universities in the United States, serving 181,000 students in a given year and offering over 800 degree options. For nine straight years, U.S. News and World Report has named ASU the most innovative university in America. Today, ASU is enhancing educational outcomes by integrating ChatGPT Edu into projects across teaching, research, and operations. Guided by the ASU charter, which prioritizes inclusion over exclusion, research benefiting the public, and responsibility for the communities they serve, ASU collaborates with OpenAI to use technology to deliver lifelong learning and drive human potential at a social scale. In the spring of 2024, ASU graduated 20,000 students—its largest class yet. “No two people learn in exactly the same way, and innovation…
61dResearch#gpt
61d ago
OpenAI o1 Contributions
OpenAI o1 Contributions Reasoning Research Foundational Contributors Ahmed El-Kishky, Daniel Selsam, Francis Song, Giambattista Parascandolo, Hongyu Ren, Hunter Lightman, Hyung Won Chung, Ilge Akkaya, Ilya Sutskever, Jason Wei, Jonathan Gordon, Karl Cobbe, Kevin Yu, Lukas Kondraciuk, Max Schwarzer, Mostafa Rohaninejad, Noam Brown, Shengjia Zhao, Trapit Bansal, Vineet Kosaraju, Wenda Zhou Leadership Jakub Pachocki, Jerry Tworek (overall), Liam Fedus, Lukasz Kaiser, Mark Chen, Szymon Sidor, Wojciech Zaremba Core Contributors Alex Karpenko, Alexander Wei, Allison Tam, Ananya Kumar, Andre Saraiva, Andrew Kondrich, Andrey Mishchenko, Ashvin Nair, Behrooz Ghorbani, Bohan Zhang, Brandon McKinzie, Brydon Eastman, Chak Ming Li, Chris Koch, Dan Roberts, David Dohan, David Mely, Dimitris Tsipras, Enoch Cheung, Eric Wallace, Hadi Salman, Haiming Bao, Hessam Bagherinezhad, Ilya Kostrikov, Jiacheng Feng, John Rizzo, Karina Nguyen, Kevin Lu, Kevin Stone, Lorenz Kuhn, Mason Meyer, Mikhail Pavlov, Nat McAleese, Oleg Boiko, Oleg Murk, Peter…
61dResearch
61d ago
Genmab launches “AI Everywhere”
Genmab launches “AI Everywhere” Genmab(opens in a new window), a leading global biotechnology company, is pioneering next-generation antibody therapies to treat cancer and other serious diseases. Their mission is ambitious: to revolutionize patient care with transformative “knock-your-socks-off” (KYSO®) antibody treatments. “Genmab’s ambition is to integrate AI into everything we do,” said Tahi Ahmadi, Executive Vice President and Chief Medical Officer, Head of Experimental Medicines. “We anticipated AI to contribute significantly to the quality of our science, decision making, and efficiency in bringing medicines to patients.” As a company that has recently tripled in size, Genmab wanted to use AI to address operational challenges—and develop new ways of working with vast amounts of complex scientific data. As part of its strategic vision to innovate and leverage AI, Genmab identified a unique opportunity to partner with ChatGPT by launching its Enterprise offering…
61d ago
Shaping the future of financial services
Morgan Stanley uses AI evals to shape the future of financial services Morgan Stanley(opens in a new window) collaborated with OpenAI to build AI solutions that empower financial advisors with faster insights, more informed decisions, and efficient summarization tools to deepen client relationships. Their success was grounded in a robust evaluation framework that ensures AI performs reliably, consistently, and at the high standards advisors expect. By embedding GPT‑4 into their workflows, Morgan Stanley Wealth Management has enhanced how financial advisors access the firm’s knowledge base and respond to client needs. Today, over 98% of advisor teams actively use AI @ Morgan Stanley Assistant—Morgan Stanley’s internal chatbot for answering financial advisors’ questions—for seamless internal information retrieval. “This technology makes you as smart as the smartest person in the organization. Each client is different, and AI helps us cater to each client’s…
63d ago
Why we no longer evaluate SWE-bench Verified
Why SWE-bench Verified no longer measures frontier coding capabilities SWE-bench Verified is increasingly contaminated. We recommend SWE-bench Pro. Since we first published SWE-bench Verified in August 2024, the industry has widely used it to measure the progress of models on autonomous software engineering tasks. After its release, SWE-bench Verified provided a strong signal of capability progress and became a standard metric reported in frontier model releases. Tracking and forecasting progress of these capabilities is also an important part of OpenAI’s Preparedness Framework. When we created the Verified benchmark initially, we attempted to solve issues in the original evaluation that made certain tasks impossible to accomplish in the SWE-bench dataset(opens in a new window). After initial leaps, state-of-the-art progress on SWE-bench Verified has slowed, improving(opens in a new window) from 74.9% to 80.9% in the last 6 months. This raises the…
63dResearch#coding#training
66d ago
Our First Proof submissions
Our First Proof submissions We’re sharing our proof attempts for First Proof, a math challenge testing if AI can produce checkable proofs on domain-specific problems. We ran an internal model on all 10 First Proof(opens in a new window) problems, a research-level math challenge designed to test whether AI systems can produce correct, checkable proof attempts. Unlike short-answer or competition-style math, these problems require building end-to-end arguments in specialized domains, and correctness is hard to establish without expert review. The authors of the First Proof problems are leading experts in their respective fields, and at least a couple of the problems were open for years before the authors found solutions. An academic department that has substantial overlap with the subject areas could conceivably solve many of the problems in one week. We shared(opens in a new window) our proof attempts…
66dResearch
67d ago
Advancing independent research on AI alignment
Advancing independent research on AI alignment We’re committing $7.5M to The Alignment Project to fund independent research developing mitigations to safety and security risks from misaligned AI. As AI systems become more capable and more autonomous, alignment research needs to both keep pace and scale diversity. At OpenAI, we invest heavily in frontier alignment and safety research as it is critical to our mission. We also believe that ensuring that AGI is safe and beneficial to everyone cannot be achieved by any single organization and want to support independent research and conceptual approaches that can be pursued outside of frontier labs. Today, we’re announcing a $7.5 million grant to The Alignment Project(opens in a new window), a global fund for independent alignment research created by the UK AI Security Institute (UK AISI). Renaissance Philanthropy is supporting the grant’s administration. This…
67dResearch#safety
68d ago
Introducing EVMbench
Introducing EVMbench Making smart contracts safer by evaluating AI agents’ ability to detect, patch, and exploit vulnerabilities in blockchain environments. Smart contracts routinely secure $100B+ in open-source crypto assets. As AI agents improve at reading, writing, and executing code, it becomes increasingly important to measure their capabilities in economically meaningful environments, and to encourage the use of AI systems defensively to audit and strengthen deployed contracts. Together with Paradigm(opens in a new window), we’re introducing EVMbench, a benchmark evaluating the ability of AI agents to detect, patch, and exploit high-severity smart contract vulnerabilities. EVMbench draws on 117 curated vulnerabilities from 40 audits, with most sourced from open code audit competitions. EVMbench additionally includes several vulnerability scenarios drawn from the security auditing process for the Tempo(opens in a new window) blockchain, a purpose-built L1 designed to enable high-throughput, low-cost payments via…
68dResearch#benchmark
73d ago
Scaling social science research
Scaling social science research A new tool to help researchers turn qualitative data into numbers they can analyze. A core part of our work at OpenAI is enabling scientists to move faster and solve harder problems. Today, our Economic Research Team is releasing GABRIEL: an open-source toolkit that uses GPT to turn unstructured text and images into quantitative measurements. It is designed for economists, social scientists, and data scientists to study qualitative data at scale. Qualitative data tells the richest stories about the world—what people say, write, teach, argue, and experience. It spans everything from syllabi and interviews to social media and photographs. There is a tremendous amount of it. But transforming that type of data into rigorous evidence is incredibly time-consuming. Often it isn't feasible at all. In too many cases, social scientists are forced to forego important avenues…
73dResearch#open-source
74d ago
Introducing GPT-5.3-Codex-Spark
Today, we’re releasing a research preview of GPT‑5.3‑Codex‑Spark, a smaller version of GPT‑5.3‑Codex, and our first model designed for real-time coding. Codex-Spark marks the first milestone in our partnership with Cerebras, which we announced in January. Codex-Spark is optimized to feel near-instant when served on ultra-low latency hardware—delivering more than 1000 tokens per second while remaining highly capable for real-world coding tasks. We’re sharing Codex-Spark on Cerebras as a research preview to ChatGPT Pro users so that developers can start experimenting early while we work with Cerebras to ramp up datacenter capacity, harden the end-to-end user experience, and deploy our larger frontier models. Our latest frontier models have shown particular strengths in their ability to do long-running tasks, working autonomously for hours, days or weeks without intervention. Codex-Spark is our first model designed specifically for working with Codex in real-time—making…
74dResearch#gpt#coding
75d ago
Harness engineering: leveraging Codex in an agent-first world
Harness engineering: leveraging Codex in an agent-first world By Ryan Lopopolo, Member of the Technical Staff Over the past five months, our team has been running an experiment: building and shipping an internal beta of a software product with 0 lines of manually-written code. The product has internal daily users and external alpha testers. It ships, deploys, breaks, and gets fixed. What’s different is that every line of code—application logic, tests, CI configuration, documentation, observability, and internal tooling—has been written by Codex. We estimate that we built this in about 1/10th the time it would have taken to write the code by hand. Humans steer. Agents execute. We intentionally chose this constraint so we would build what was necessary to increase engineering velocity by orders of magnitude. We had weeks to ship what ended up being a million lines of…
81d ago
GPT-5 lowers the cost of cell-free protein synthesis
GPT‑5 lowers the cost of cell-free protein synthesis Working with Ginkgo Bioworks, we created an AI-driven autonomous lab and achieved a 40% reduction in protein production cost. We’ve seen rapid progress from AI in fields like math and physics, where ideas can often be evaluated without touching the physical world. Biology is different. Progress runs through the lab, where scientists run experiments that take time and money. That’s starting to change. Frontier models can now connect directly to lab automation, propose experiments, run them at scale, learn from the results, and decide what to do next. In much of life science, the bottleneck is iteration, and autonomous labs are built to remove that constraint. In earlier work, we showed that GPT‑5 could improve wet-lab protocols through closed-loop experimentation. Here, we show that the same approach can reduce the cost of…
81dResearch#agents
89d ago
EMEA Youth & Wellbeing Grant
EMEA Youth & Wellbeing Grant Supporting organizations improving youth safety and wellbeing in the age of AI. April 2026 update: Applications for 2026 are now closed. Finalists for the EMEA Youth & Wellbeing Grant program have now been selected. If you have not been contacted by the team, it unfortunately means your project was not selected for the final round. We truly appreciate the time and effort invested in your application and the important work you are doing. The EMEA Youth & Wellbeing Grant is a €500,000 funding program for organizations across the region to help young people benefit from AI. The program will offer funding to NGOs and research organizations that are working directly with children, young people, families or educators; or producing independent research on how AI affects young people’s safety, wellbeing and development. Our goal is to…
89dResearch#safety
89d ago
The next chapter for AI in the EU
Key takeaways: - New program to train 20,000 SME across Europe with AI skills - €500,000 NGO grant to support research into youth safety and wellbeing - More ways to partner with governments through OpenAI for Europe OpenAI is today launching its EU Economic Blueprint 2.0—with new EU AI usage data and a set of initiatives designed to accelerate adoption of AI across Europe to ensure people, businesses and countries seize the full opportunity of this transformative technology. That includes a new program to train 20,000 SMEs across Europe with AI skills, supported by Booking.com; a €500,000 NGO grant to support EU research into youth safety and wellbeing; and more ways for governments to partner with OpenAI on national AI priorities through the OpenAI for Europe initiative. The Blueprint shares new data from OpenAI on Europe’s growing AI capability overhang—the…
89dResearch#safety
90d ago
Introducing Prism
Introducing Prism Accelerating science writing and collaboration with AI. Science shapes nearly every part of daily life—from the medicines we rely on, to the energy that powers our homes, to the systems that keep us safe. But the pace of scientific progress is still constrained by how research is done day to day. While AI has advanced rapidly, much of the everyday work of science still relies on tools that haven’t fundamentally changed in decades. We’re introducing Prism, a free, AI-native workspace for scientists to write and collaborate on research, powered by GPT‑5.2. Prism offers unlimited projects and collaborators and is available today to anyone with a ChatGPT personal account. Prism will be available soon to organizations using ChatGPT Business, Enterprise, and Education plans. Over the past year, we’ve begun to see AI accelerate scientific work across domains. Advanced reasoning…
90dResearch
[PB]PyTorch Blog· 2 articlesvisit →
19d ago
SOTA Normalization Performance with torch.compile
Introduction Normalization methods (LayerNorm/RMSNorm) are foundational in deep learning and are used to normalize values of inputs to result in a smoother training process for deep learning models. We evaluate and improve torch.compile performance for LayerNorm/RMSNorm on NVIDIA H100 and B200 to reach near SOTA performance on a kernel-by-kernel basis, in addition with further speedups through automatic fusion capabilities. Forwards LayerNorm LayerNorm was first introduced in this paper: https://arxiv.org/abs/1607.06450. It normalizes the inputs by taking the mean and variance, along with scaling by learnable parameters, gamma (weight) and Beta (bias). RMSNorm RMSNorm (root mean square norm) was introduced as a follow up of LayerNorm in this paper: https://arxiv.org/abs/1910.07467. Instead of centering on the mean, the RMS is used to normalize, which is a sum of the squares of x values. We still use gamma (weight) as a learnable parameter for…
19dResearch#training#gpu#safetyby Shunting Zhang, Paul Zhang, Elias Ellison, Markus Hoehnerbach, Jason Ansel, Natalia Gimelshein
33d ago
Flight Recorder: A New Lens for Understanding NCCL Watchdog Timeouts
If you’ve ever trained a large AI model and had it fail with an error like: [Rank 0] Watchdog caught collective operation timeout: WorkNCCL(SeqNum=12345, OpType=ALLREDUCE, NumelIn=1, NumelOut=1, Timeout(ms)=600000) ran for 600029 milliseconds before timing out. Exception raised from checkTimeout at .../torch/csrc/distributed/c10d/ProcessGroupNCCL.cpp:692 (most recent call first): ... # 2 c10d::ProcessGroupNCCL::WorkNCCL::checkTimeout(std::optional<std::chrono::duration<long, std::ratio<1l, 1000l> > >) # 3 c10d::ProcessGroupNCCL::Watchdog::runLoop() # 4 c10d::ProcessGroupNCCL::Watchdog::run() # 5 execute_native_thread_routine # 6 start_thread # 7 __clone3 You’ve encountered the infamous NCCL watchdog timeout. Debugging this error can be hard – the error message is generic, debugging requires cross-rank telemetry analysis, and root causes are multi-layered and can have a complex causal chain. This post provides key insights on NCCL watchdog timeouts, including: - Why this error happens and why it’s so hard to debug; - A deep dive into the most common root causes for the error (e.g.,…
33dResearchby Phillip Liu, Uttam Thakore, Junjie Wang, Justin Yang
[SWB]Simon Willison Blog· 4 articlesvisit →
5d ago
Quoting Bobby Holley
22nd April 2026 As part of our continued collaboration with Anthropic, we had the opportunity to apply an early version of Claude Mythos Preview to Firefox. This week’s release of Firefox 150 includes fixes for 271 vulnerabilities identified during this initial evaluation. [...] Our experience is a hopeful one for teams who shake off the vertigo and get to work. You may need to reprioritize everything else to bring relentless and single-minded focus to the task, but there is light at the end of the tunnel. We are extremely proud of how our team rose to meet this challenge, and others will too. Our work isn’t finished, but we’ve turned the corner and can glimpse a future much better than just keeping up. Defenders finally have a chance to win, decisively. — Bobby Holley, CTO, Firefox Recent articles - DeepSeek…
5dResearch#claude
5d ago
Qwen3.6-27B: Flagship-Level Coding in a 27B Dense Model
22nd April 2026 - Link Blog Qwen3.6-27B: Flagship-Level Coding in a 27B Dense Model (via) Big claims from Qwen about their latest open weight model: Qwen3.6-27B delivers flagship-level agentic coding performance, surpassing the previous-generation open-source flagship Qwen3.5-397B-A17B (397B total / 17B active MoE) across all major coding benchmarks. On Hugging Face Qwen3.5-397B-A17B is 807GB, this new Qwen3.6-27B is 55.6GB. I tried it out with the 16.8GB Unsloth Qwen3.6-27B-GGUF:Q4_K_M quantized version and llama-server using this recipe by benob on Hacker News, after first installing llama-server using brew install llama.cpp : llama-server \ -hf unsloth/Qwen3.6-27B-GGUF:Q4_K_M \ --no-mmproj \ --fit on \ -np 1 \ -c 65536 \ --cache-ram 4096 -ctxcp 2 \ --jinja \ --temp 0.6 \ --top-p 0.95 \ --top-k 20 \ --min-p 0.0 \ --presence-penalty 0.0 \ --repeat-penalty 1.0 \ --reasoning on \ --chat-template-kwargs '{"preserve_thinking": true}' On first run that…
7d ago
SQL functions in Google Sheets to fetch data from Datasette
20th April 2026 TIL SQL functions in Google Sheets to fetch data from Datasette — I've been experimenting with ways to fetch data from Datasette and display it in Google Sheets. I put together some notes on patterns for fetching data from a Datasette instance directly into Google Sheets - using the importdata() function, a "named function" that wraps it or a Google Apps Script if you need to send an API token in an HTTP header (not supported by importdata() .) Here's an example sheet demonstrating all three methods. Recent articles - DeepSeek V4 - almost on the frontier, a fraction of the price - 24th April 2026 - Extract PDF text in your browser with LiteParse for the web - 23rd April 2026 - A pelican for GPT-5.5 via the semi-official Codex backdoor API - 23rd April 2026
7dResearch
9d ago
Claude system prompts as a git timeline
18th April 2026 Research Claude system prompts as a git timeline — Anthropic's published system prompt history for Claude is transformed into a git-based exploration tool, breaking up the monolithic markdown source into granular files and timestamped commits. By structuring extracted prompts per model, family, and revision, researchers can leverage `git log`, `diff`, and `blame` to trace prompt evolution, compare differences, and attribute changes to specific dates—all without manual parsing. Anthropic publish the system prompts for Claude chat and make that page available as Markdown. I had Claude Code turn that page into separate files for each model and model family with fake git commit dates to enable browsing the changes via the GitHub commit view. I used this to write my own detailed notes on the changes between Opus 4.6 and 4.7. Recent articles - DeepSeek V4 - almost…
9dResearch#claude#coding
[TG]The Gradient· 1 articlesvisit →
68d ago
After Orthogonality: Virtue-Ethical Agency and AI Alignment
Preface This essay argues that rational people don’t have goals, and that rational AIs shouldn’t have goals. Human actions are rational not because we direct them at some final ‘goals,’ but because we align actions to practices[1]: networks of actions, action-dispositions, action-evaluation criteria, and action-resources that structure, clarify, develop, and promote themselves. If we want AIs that can genuinely support, collaborate with, or even comply with human agency, AI agents’ deliberations must share a “type signature” with the practices-based logic we use to reflect and act. I argue that these issues matter not just for aligning AI to grand ethical ideals like human flourishing, but also for aligning AI to core safety-properties like transparency, helpfulness, harmlessness, or corrigibility. Concepts like ’harmlessness’ or ‘corrigibility’ are unnatural -- brittle, unstable, arbitrary -- for agents who’d interpret them in terms of goals or…
68dResearch#safetyby Peli Grietzer
[TVA]The Verge AI· 1 articlesvisit →
3d ago
How Project Maven taught the military to love AI
In the first 24 hours of the assault on Iran, the US military struck more than 1,000 targets, nearly double the scale of the “shock and awe” attack on Iraq over two decades ago. This acceleration was made possible by AI systems that speed up the targeting process. Chief among them is the Maven Smart System. How Project Maven taught the military to love AI A new book shows how the controversial Silicon Valley partnership has accelerated the pace of war How Project Maven taught the military to love AI A new book shows how the controversial Silicon Valley partnership has accelerated the pace of war In her new book, Project Maven: A Marine Colonel, His Team, and the Dawn of AI Warfare, journalist Katrina Manson investigates the development of Maven from its inception in 2017 as an experiment in…
3dResearchby Joshua Dzieza
[WA]Wired AI· 3 articlesvisit →
2d ago
Discord Sleuths Gained Unauthorized Access to Anthropic’s Mythos
As researchers and practitioners debate the impact that new AI models will have on cybersecurity, Mozilla said on Tuesday it used early access to Anthropic's Mythos Preview to find and fix 271 vulnerabilities in its new Firefox 150 browser release. Meanwhile, researchers identified a group of moderately successful North Korean hackers using AI for everything from vibe coding malware to creating fake company websites—stealing up to $12 million in three months. Researchers have finally cracked disruptive malware known as Fast16 that predates Stuxnet and may have been used to target Iran’s nuclear program. It was created in 2005 and was likely deployed by the US or an ally. Meta is being sued by the Consumer Federation of America, a nonprofit, over scam ads on Facebook and Instagram and allegedly misleading consumers about the company’s efforts to combat them. A United…
2dResearch#codingby Matt Burgess, Lily Hay Newman, Andy Greenberg
3d ago
Apple's Next CEO Needs to Launch a Killer AI Product
Sometime in the next year or two, Apple’s new CEO, John Ternus, will step onto a stage and tell the world that his company has a revolutionary product. This product, he’ll say, will put the full and awesome power of AI into everyone’s hands. It probably won’t represent a breakthrough in AI research, and it might not let people automate work or perform tasks any better than a lot of technically minded people are doing today. It may or may not involve a new device, though if it doesn’t, one should be in development. But if it all works out, that keynote will mark the moment when Apple did to AI what it has done for desktop computers, the internet, mobile technology, wearables, and music distribution. That is, it’ll offer a solution to a troublesome technology that’s so delightful and…
3dResearchby Steven Levy
3d ago
Ace the Ping-Pong Robot Can Whup Your Ass
Ace is a robot that aims high: It wants to become the world champion of table tennis. It was developed by Sony AI researchers who, in a new study published in Nature, have shown how this robot, equipped with artificial intelligence, has faced some high-level athletes, holding its own in matches played according to the official rules of table tennis. This feat represents a milestone for the world of robotics, a field that has long regarded this sport, among the most technical in the world, as one of the most difficult tests of technological advances. Robot Player We have already seen artificial intelligence systems win virtual competitions in games such as chess, Go, and even StarCraft II, but physical games are much more difficult to master. A robot needs to sense unpredictable changes in the external environment, interpret their meaning,…
3dResearchby Marta Musso