<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:media="http://search.yahoo.com/mrss/"><channel><title>IEEE Spectrum</title><link>https://spectrum.ieee.org/</link><description>IEEE Spectrum</description><atom:link href="https://spectrum.ieee.org/feeds/topic/artificial-intelligence.rss" rel="self"></atom:link><language>en-us</language><lastBuildDate>Wed, 15 Apr 2026 20:00:41 -0000</lastBuildDate><image><url>https://spectrum.ieee.org/media-library/eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpbWFnZSI6Imh0dHBzOi8vYXNzZXRzLnJibC5tcy8yNjg4NDUyMC9vcmlnaW4ucG5nIiwiZXhwaXJlc19hdCI6MTgyNjE0MzQzOX0.N7fHdky-KEYicEarB5Y-YGrry7baoW61oxUszI23GV4/image.png?width=210</url><link>https://spectrum.ieee.org/</link><title>IEEE Spectrum</title></image><item><title>​Boston Dynamics and Google DeepMind Teach Spot to Reason​</title><link>https://spectrum.ieee.org/boston-dynamics-spot-google-deepmind</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/photo-of-yellow-boston-dynamics-robot-dog-using-its-arm-to-load-laundry-into-a-white-basket.png?id=65521323&width=1200&height=800&coordinates=150%2C0%2C150%2C0"/><br/><br/><p><span><strong></strong><strong></strong>The amazing and frustrating thing about robots is that they can do almost anything you want them to do, as long as you know how to ask properly. In the not-so-distant past, asking properly meant writing code, and while we’ve thankfully moved beyond that brittle constraint, there’s still an irritatingly inverse correlation between ease of use and complexity of task. </span></p><p><span>AI has promised to change that. The idea is that when AI is embodied within robots—giving AI software a physical presence in the world—those robots will be imbued with reasoning and understanding. This is cutting-edge stuff, though, and while we’ve seen plenty of examples of embodied AI in a research context, finding applications where reasoning robots can provide reliable commercial value has not been easy. <a href="https://bostondynamics.com/" target="_blank">Boston Dynamics</a> is one of the few companies to commercially deploy legged robots at any appreciable scale; there are now several thousand hard at work. Today the company is <a href="https://bostondynamics.com/blog/tools-for-your-to-do-list-with-spot-and-gemini-robotics/" target="_blank">announcing</a> that its quadruped robot <a href="https://spectrum.ieee.org/tag/spot-robot" target="_self">Spot</a> is now equipped with <a href="https://deepmind.google/blog/gemini-robotics-er-1-6/">Google DeepMind’s Gemini Robotics-ER 1.6</a>, a <a href="https://spectrum.ieee.org/gemini-robotics" target="_blank">high-level embodied reasoning model</a> that brings usability and intelligence to complex tasks.</span></p><p class="shortcode-media shortcode-media-youtube"> <span class="rm-shortcode" data-rm-shortcode-id="155eddc016bd1bedcfb5b83c4b4a54c3" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/LP4-c5AK30g?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">YouTube.com</small></p><p><span>Although this video shows Spot in a home context, the focus of this partnership is on one of the very few applications where legged robots have proven themselves to be commercially viable: inspection. That is, wandering around industrial facilities, checking to make sure that nothing is imminently exploding. With the new AI onboard, Spot is now able to autonomously look for dangerous debris or spills, read complex gauges and sight glasses, and call on tools like vision-language-action models when it needs help understanding what’s going on in the environment around it.</span></p><p>“Advances like Gemini Robotics-ER 1.6 mark an important step toward robots that can better understand and operate in the physical world,” <a href="https://www.linkedin.com/in/marco-da-silva-447b72/" target="_blank">Marco da Silva</a>, vice president and general manager of Spot at Boston Dynamics, says <a href="https://bostondynamics.com/blog/aivi-learning-now-powered-google-gemini-robotics/" target="_blank">in a press release</a>. “Capabilities like instrument reading and more reliable task reasoning will enable Spot to see, understand, and react to real-world challenges completely autonomously.”</p><h2>Understanding Robot Understanding</h2><p>The words “reasoning” and “understanding” are being increasingly applied to AI and robotics, but as <a href="https://spectrum.ieee.org/humanoid-robots-gill-pratt-darpa" target="_self">Toyota Research Institute’s Gill Pratt recently pointed out</a>, what those words actually <em><em>mean</em></em> for robots in practice isn’t always clear. “The benchmark we measure ourselves against when it comes to understanding is that the system should answer the way a human would,” <a href="https://www.linkedin.com/in/carolinaparada/" target="_blank">Carolina Parada</a>, head of robotics at Google DeepMind, explained in an interview. For robots to reliably and safely perform tasks, this connection between how robots understand the world and how humans do is critical. Otherwise, there may be a disconnect between the instructions that a human gives a robot, and how the robot decides to carry out that task.</p><p>Boston Dynamics’ video above is a potentially messy example of this. One of the instructions to Spot was to “recycle any cans in the living room.” It has no problem completing the task, as the video shows, but in doing so, it grips the can sideways, which is not going to end up well for cans that have leftover liquid in them. We humans would avoid this because we can draw on a lifetime of experience to know how cans should be held, but robots don’t (yet) have that kind of world knowledge.</p><p>Parada says that Gemini Robotics-ER 1.6 approaches situations like this from a safety perspective. “If you ask the robot to bring you a cup of water, it will reason not to place it on the edge of a table where it could fall. We track this using our <a href="https://asimov-benchmark.github.io/v1/" target="_blank">ASIMOV benchmark</a>, which includes a whole lot of natural language examples of things the robot should not do.” The current version of Spot doesn’t use these semantic safety models for manipulation, but the plan is to make future versions reason about holding objects in ways that are safe.</p><p class="shortcode-media shortcode-media-youtube" style="background-color: rgb(255, 255, 255);"><span class="rm-shortcode" data-rm-shortcode-id="5934a9a019325c2e996f3f0dab47b3c4" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/kBwxmlI2yHQ?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">YouTube.com</small></p><p><span>There does still seem to be a disconnect between Gemini Robotics-ER 1.6 as a high-level reasoning model for a robot, and the robot itself as an interface with the physical world. One of the new features of 1.6 is </span><em><em>success detection</em></em><span>, which combines multiple camera angles to more reliably be able to tell when Spot has successfully grasped an object. This is great if you’re relying entirely on vision for your object interaction, but robots have all kinds of other well-established ways to detect a successful grasp, including touch sensors and force sensors, that 1.6 is not using. The reason why this is the case speaks to a fundamental problem that the robotics field is still trying to figure out: how to train models when you need physical data.</span></p><p><span>“At the moment, these models are strictly vision only,” Parada explains. “There is lots of [visual] information on the web about how to pick up a pen. If we had enough data with touch information, we could easily learn it, but there is not a lot of data with touch sensing on the internet.” Customers who use these new capabilities for inspection with Spot will be required to share their data with Boston Dynamics, which is where some of this data will come from.</span></p><h2>Real-World Robots That Are Useful</h2><p>The fact that Boston Dynamics <em><em>has </em></em>customers makes them something of an anomaly when it comes to legged robots that rely on AI in commercial deployments. And those customers will have to be able to trust the robot—<a href="https://spectrum.ieee.org/ai-hallucination" target="_self">always a problem when AI is involved</a>. “We take this very seriously,” da Silva said in an interview. “We roll out new DeepMind capabilities through beta programs to a smaller set of customers to understand what to anticipate, and we only actively advertise features we are confident will work.” There’s a threshold of usefulness that robots like Spot need to reach, and fortunately, the real world doesn’t demand perfection. “Most critical infrastructure in a facility will be instrumented to tell you whether something is wrong,” da Silva says. “But there is a lot of stuff that is not instrumented that can still cause a problem if you aren’t paying attention to it. We’ve found that somewhere north of 80 percent is the threshold where it’s not annoying. Below that, basically the robot is crying wolf, and the operators will start ignoring it.”</p><p><span></span><span>Both da Silva and Parada agree that there’s still plenty of room for improvement in robotic inspection. As Parada points out, Spot’s rarefied status as a scalable commercial platform provides a valuable opportunity to learn how models like Gemini Robotics-ER 1.6 can be the most useful, and then apply that knowledge to other embodied AI platforms, including </span><a href="https://spectrum.ieee.org/boston-dynamics-atlas-scott-kuindersma" target="_self">Boston Dynamics’ Atlas</a><span>. Does that mean that Atlas is going to be the next industrial inspection robot? Probably not. But if this real-world experience can get us closer to safe and reliable robots that can pick up laundry, take a dog for a walk, and clear away soda cans without making a mess, that’s something we can all get excited about.</span></p>]]></description><pubDate>Tue, 14 Apr 2026 19:45:01 +0000</pubDate><guid>https://spectrum.ieee.org/boston-dynamics-spot-google-deepmind</guid><category>Boston-dynamics</category><category>Spot-robot</category><category>Google-deepmind</category><category>Inspection-robots</category><category>Quadruped-robots</category><dc:creator>Evan Ackerman</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/photo-of-yellow-boston-dynamics-robot-dog-using-its-arm-to-load-laundry-into-a-white-basket.png?id=65521323&amp;width=980"></media:content></item><item><title>Sarang Gupta Builds AI Systems With Real-World Impact</title><link>https://spectrum.ieee.org/openai-engineer-sarang-gupta</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-young-adult-indian-man-smiling-with-his-arms-crossed.png?id=65519413&width=1200&height=800&coordinates=0%2C83%2C0%2C84"/><br/><br/><p>Like many engineers, <a href="https://www.linkedin.com/in/sarang-gupta/" rel="noopener noreferrer" target="_blank">Sarang Gupta</a> spent his childhood tinkering with everyday items around the house. From a young age he gravitated to projects that could make a difference in someone’s everyday life.</p><p>When the family’s microwave plug broke, Gupta and his father figured out how to fix it. When a drawer handle started jiggling annoyingly, the youngster made sure it didn’t do so for long.</p><h3>Sarang Gupta</h3><br/><p><strong>Employer</strong></p><p><strong></strong>OpenAI in San Francisco</p><p><strong>Job</strong></p><p><strong></strong>Data science staff member</p><p><strong>Member grade</strong></p><p>Senior member</p><p><strong>Alma maters </strong></p><p><strong></strong>The Hong Kong University of Science and Technology; Columbia</p><p>By age 11, his interest expanded from nuts and bolts to software. He learned <a data-linked-post="2674010559" href="https://spectrum.ieee.org/top-programming-languages-2025" target="_blank">programming languages</a> such as <a href="https://en.wikipedia.org/wiki/BASIC" rel="noopener noreferrer" target="_blank">Basic</a> and <a href="https://en.wikipedia.org/wiki/Logo_(programming_language)" rel="noopener noreferrer" target="_blank">Logo</a> and designed simple programs including one that helped a local restaurant automate online ordering and billing.</p><p>Gupta, an IEEE senior member, brings his mix of curiosity, hands-on problem-solving, and a desire to make things work better to his role as member of the data science staff at <a href="https://openai.com/" rel="noopener noreferrer" target="_blank">OpenAI</a> in San Francisco. He works with the go-to-market (GTM) team to help businesses adopt <a href="https://chatgpt.com/" rel="noopener noreferrer" target="_blank">ChatGPT</a> and other products. He builds data-driven models and systems that support the sales and marketing divisions.</p><p>Gupta says he tries to ensure his work has an impact. When making decisions about his career, he says, he thinks about what AI solutions he can unlock to improve people’s lives.</p><p>“If I were to sum up my overall goal in one sentence,” he says, “it’s that I want AI’s benefits to reach as many people as possible.”</p><h2>Pursuing engineering through a business lens</h2><p>Gupta’s early interest in tinkering and programming led him to choose physics, chemistry, and math as his higher-level subjects at <a href="https://www.cirschool.org/" rel="noopener noreferrer" target="_blank">Chinmaya International Residential School</a>, in Tamil Nadu, India. As part of the high school’s <a href="https://www.ibo.org/" rel="noopener noreferrer" target="_blank">International Baccalaureate</a> chapter, students select three subjects in which to specialize.</p><p>“I was interested in engineering, including the theoretical part of it,” Gupta says, “But I was always more interested in the applications: how to sell that technology or how it ties to the real world.”</p><p>After graduating in 2012, he moved overseas to attend the <a href="https://hkust.edu.hk/" rel="noopener noreferrer" target="_blank">Hong Kong University of Science and Technology</a>. The university offered a <a href="https://techmgmt.hkust.edu.hk/" rel="noopener noreferrer" target="_blank">dual bachelor’s program</a> that allowed him to earn one degree in industrial engineering and another in business management in just four years.</p><p>In his spare time, Gupta built a smartphone app that let students upload their class schedules and find classmates to eat lunch with. The app didn’t take off, he says, but he enjoyed developing it. He also launched Pulp Ads, a business that printed advertisements for student groups on tissues and paper napkins, which were distributed in the school’s cafeterias. He made some money, he says, but shuttered the business after about a year.</p><p>After graduating from the university in 2016, he decided to work in Hong Kong’s financial hub and joined <a href="https://www.goldmansachs.com/" rel="noopener noreferrer" target="_blank">Goldman Sachs</a> as an analyst in the bank’s operations division.</p><h2>From finance to process optimization at scale</h2><p>After two parties agree on securities transactions, the bank’s operations division ensures that the trade details are recorded correctly, the securities and payments are ready to transfer, and the transaction settles accurately and on time.</p><p>As an analyst, Gupta’s task was to find bottlenecks in the bank’s workflows and fix them. He identified an opportunity to automate trade reconciliation: when analysts would manually compare data across spreadsheets and systems to make sure a transaction’s details were consistent. The process helped ensure financial transactions were recorded accurately and settled correctly.</p><p>Gupta built internal automation tools that pulled trade data from different systems, ran validation checks, and generated reports highlighting any discrepancies.</p><p>“Instead of analysts manually checking large datasets, the tools automatically flagged only the cases that required investigation,” he says. “This helped the team spend less time on repetitive verification tasks and more time resolving complex issues. It was also my first real exposure to how software and data systems could dramatically improve operational workflows.”</p><p class="pull-quote">“Whether it’s helping a person improve a trait like that or driving efficiencies at a business, AI just has so much potential to help. I’m excited to be a little part of that.”</p><p>The experience made him realize he wanted to work more deeply in technology and data-driven systems, he says. He decided to return to school in 2018 to study data science and AI, when the fields were just beginning to surge into broader awareness.</p><p>He discovered that <a href="https://www.columbia.edu/" rel="noopener noreferrer" target="_blank">Columbia</a> offered a dedicated master’s degree program in data science with a focus on AI. After being accepted in 2019, he moved to New York City.</p><p>Throughout the program, he gravitated to the applied side of machine learning, taking courses in applied deep learning and neural networks.</p><p>One of his major academic highlights, he says, was a project he did in 2019 with the <a href="https://brown.columbia.edu/" rel="noopener noreferrer" target="_blank">Brown Institute</a>, a joint research lab between Columbia and <a href="https://www.stanford.edu/" rel="noopener noreferrer" target="_blank">Stanford</a> focused on using technology to improve journalism. The team worked with <a href="https://www.inquirer.com/" rel="noopener noreferrer" target="_blank"><em><em>The Philadelphia Inquirer</em></em></a><em> </em>to help the newsroom staff better understand their coverage from a geographic and social standpoint. The project highlighted “news deserts”—underserved communities for which the newspaper was not providing much coverage—so the publication could redirect its reporting resources.</p><p>To identify those areas, <a href="https://aclanthology.org/2020.nlpcss-1.17.pdf" rel="noopener noreferrer" target="_blank">Gupta and his team built tools that extracted locations such as</a> street names and neighborhoods from news articles and mapped them to visualize where most of the coverage was concentrated. The <em><em>Inquirer</em></em> implemented the tool in several ways including a new <a href="https://medium.com/the-lenfest-local-lab/how-we-built-a-tool-to-spot-geographic-clusters-and-gaps-in-local-news-e553abe88287" rel="noopener noreferrer" target="_blank">web page that aggregated stories about COVID-19 by county</a>.</p><p> “Journalism was an interesting problem set for me, because I really like to read the news every day,” Gupta says. “It was an opportunity to work with a real newsroom on a problem that felt really impactful for both the business and the local community.”</p><h2>The GenAI inflection point</h2><p>After earning his master’s degree in 2020, Gupta moved to San Francisco to join <a href="https://asana.com/" rel="noopener noreferrer" target="_blank">Asana</a>, the company that developed the work management platform by the same name. He was drawn to the opportunity to work for a relatively small company where he could have end-to-end ownership of projects. He joined the organization as a product data scientist, focusing on A/B testing for new platform features.</p><p>Two years later, a new opportunity emerged: He was asked to lead the launch of Asana Intelligence, an internal machine learning team building AI-powered features into the company’s products.</p><p>“I felt I didn’t have enough experience to be the founding data scientist,” he says. “But I was also really interested in the space, and spinning up a whole machine learning program was an opportunity I couldn’t turn down.”</p><p>The Asana Intelligence team was given six months to build several machine learning–powered features to help customers work more efficiently. They included automatic summaries of project updates, insights about potential risks or delays, and recommendations for next steps.</p><p>The team met that goal and launched several other features including <a href="https://help.asana.com/s/article/smart-status" target="_blank">Smart Status</a>, an AI tool that analyzes a project’s tasks, deadlines, and activity, then generates a status update.</p><p>“When you finally launch the thing you’ve been working on, and you see the usage go up, it’s exhilarating,” he says. “You feel like that’s what you were building toward: users actually seeing and benefiting from what you made.”</p><p>Gupta and his team also translated that first wave of work into reusable frameworks and documentation to make it easier to create machine learning features at Asana. He and his colleagues filed several <a href="https://patents.google.com/patent/US20250355685A1/" rel="noopener noreferrer" target="_blank">U.S. patents</a>.</p><p>At the time he took on that role, OpenAI launched ChatGPT. The mainstreaming of generative AI and large language models shifted much of his work at Asana from model development to assessing LLMs.</p><p>OpenAI captured the attention of people around the world, including Gupta. In September 2025 he left Asana to join OpenAI’s data science team.</p><p>The transition has been both energizing and humbling, he says. At OpenAI, he works closely with the marketing team to help guide strategic decisions. His work focuses on developing models to understand the efficiency of different marketing channels, to measure what’s driving impact, and to help the company better reach and serve its customers.</p><p>“The pace is very different from my previous work. Things move quickly,” he says. “The industry is extremely competitive, and there’s a strong expectation to deliver fast. It’s been a great learning experience.”</p><p>Gupta says he plans to stay in the AI space. With technology evolving so rapidly, he says, he sees enormous potential for task automation across industries. AI has already transformed his core software engineering work, he says, and it’s helped him enhance areas that aren’t natural strengths.</p><p>“I’m not a good writer, and AI has been huge in helping me frame my words better and <a href="https://spectrum.ieee.org/engineering-communication" target="_blank">present my work more clearly</a>,” he says. “Whether it’s helping a person improve a trait like that or driving efficiencies at a business, AI just has so much potential to help. I’m excited to be a little part of that.”</p><h2>Exploring IEEE publications and connections</h2><p>Gupta has been an IEEE member since 2024, and he values the organization as both a technical resource and a professional network.</p><p>He regularly turns to IEEE publications and the <a href="https://ieeexplore.ieee.org/Xplore/guesthome.jsp" rel="noopener noreferrer" target="_blank">IEEE Xplore Digital Library</a> to read articles that keep him abreast of the evolution of AI, data science, and the engineering profession.</p><p>IEEE’s <a href="https://cis.ieee.org/activities/membership-activities/ieee-member-directory" rel="noopener noreferrer" target="_blank">member directory</a> tools are another valuable resource that he uses often, he says.</p><p>“It’s been a great way to connect with other engineers in the same or similar fields,” he says. “I love sharing and hearing about what folks are working on. It brings me outside of what I’m doing day to day.</p><p>“It inspires me, and it’s something I really enjoy and cherish.”</p>]]></description><pubDate>Tue, 14 Apr 2026 18:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/openai-engineer-sarang-gupta</guid><category>Ieee-member-news</category><category>Openai</category><category>Generative-ai</category><category>Chatgpt</category><category>Careers</category><category>Type-ti</category><dc:creator>Julianne Pepitone</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/a-young-adult-indian-man-smiling-with-his-arms-crossed.png?id=65519413&amp;width=980"></media:content></item><item><title>12 Graphs That Explain the State of AI in 2026</title><link>https://spectrum.ieee.org/state-of-ai-index-2026</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/squares-and-rectangles-on-graph-paper-form-the-letters-ai.jpg?id=65506010&width=1200&height=800&coordinates=0%2C83%2C0%2C84"/><br/><br/><p>The capabilities of leading AI models continue to accelerate, and the largest AI companies, including <a href="https://www.cnbc.com/2026/04/08/openai-ipo-sarah-friar-retail-investors.html" target="_blank">OpenAI</a> and <a href="https://fortune.com/2026/04/10/anthropic-too-dangerous-to-release-ai-model-means-for-its-upcoming-ipo/" target="_blank">Anthropic</a>, are hurtling toward IPOs later this year. Yet resentment toward AI continues to simmer, and in some cases has boiled over, especially in the United States, where local governments are beginning to embrace restrictions or outright bans on new data center development.</p><p>It’s a lot to keep track of, but the 2026 edition of the <a href="https://hai.stanford.edu/ai-index" target="_blank">AI Index</a> from Stanford University’s <a href="https://hai.stanford.edu/" target="_blank">Human-Centered Artificial Intelligence</a> center pulls it off. The report, which comes in at over 400 pages, includes dozens of data points and graphs that approach the topic from multiple angles, from benchmark scores to investment and public perception. <br/><br/>As in prior years (see our coverage from <a href="https://spectrum.ieee.org/the-state-of-ai-in-15-graphs" target="_self">2021</a>, <a href="https://spectrum.ieee.org/artificial-intelligence-index" target="_self">2022</a>, <a href="https://spectrum.ieee.org/state-of-ai-2023" target="_self">2023</a>, <a href="https://spectrum.ieee.org/ai-index-2024" target="_self">2024</a>, and <a href="https://spectrum.ieee.org/ai-index-2025" target="_self">2025</a>), we’ve read the report and identified the trends that encapsulate the state of AI in 2026.</p><h2>US companies lead in AI models</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Chart showing the number of AI models in the United States, China, and Europe as rising from 2005 to 2025, particularly in China at 30 and the United States at 50, while Europe is at 2." class="rm-shortcode" data-rm-shortcode-id="159dac45bd56ec14147f84e06df793b1" data-rm-shortcode-name="rebelmouse-image" id="04bdb" loading="lazy" src="https://spectrum.ieee.org/media-library/chart-showing-the-number-of-ai-models-in-the-united-states-china-and-europe-as-rising-from-2005-to-2025-particularly-in-china.jpg?id=65506052&width=980"/> </p><p><span>The United States has led the charge in AI model releases over the past decade, and that remains as true in 2025 as in any year prior. According to research institute Epoch AI, organizations based in the United States released 50 “notable” models in 2025. However, China’s output is beginning to close the gap.</span></p><p>Nearly all the notable models originated within industry (as opposed to academic or government institutions). Epoch AI tracked 87 notable model releases from industry in 2025, compared to just seven from all other sources. This is a major long-term trend. Models released by industry now make up over 90 percent of notable models, up from just under 50 percent in 2015, and zero in 2003.</p><h2>China leads in robotics</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A line chart of the number of new industrial robots installed in Germany, South Korea, the United States, Japan, and China showing a massive amount more in China." class="rm-shortcode" data-rm-shortcode-id="c8eab0a1f3e6cee9558f0fd235a0b912" data-rm-shortcode-name="rebelmouse-image" id="ab17d" loading="lazy" src="https://spectrum.ieee.org/media-library/a-line-chart-of-the-number-of-new-industrial-robots-installed-in-germany-south-korea-the-united-states-japan-and-china-showi.jpg?id=65506352&width=980"/> </p><p>While U.S. companies released the largest number of notable AI models, China has an equally clear lead in the deployment of robotics. According to data from the International Federation of Robotics, China installed 295,000 industrial robots in 2024. Japan installed roughly 44,500, and the United States installed 34,200.</p><h2>World AI compute capacity has grown 3.3x yearly since 2022</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Bar chart showing the portion of global computing capacity from AI chips from Nvidia, Google, Amazon, AMD and Huawei, mostly dominated by Nvidia." class="rm-shortcode" data-rm-shortcode-id="29c8f6c52d2ab01a3f69f6627314e90f" data-rm-shortcode-name="rebelmouse-image" id="0a5a1" loading="lazy" src="https://spectrum.ieee.org/media-library/bar-chart-showing-the-portion-of-global-computing-capacity-from-ai-chips-from-nvidia-google-amazon-amd-and-huawei-mostly-dom.jpg?id=65506056&width=980"/> </p><p><span>The latest Stanford AI Index report has no shortage of head-turning numbers on the AI build-out, but none beats EpochAI’s gauge of total AI compute.</span></p><p>This graph, which uses the compute power of Nvidia’s H100e as a yardstick, shows that the world’s AI compute capacity has increased more than threefold every year since 2022. Total AI compute has increased 30-fold since 2021, the first year tracked. </p><p>Nvidia has benefited most from this build-out, as its GPUs account for over 60 percent of the total AI compute capacity in the world today. Amazon and Google—each of which design their own hardware for AI workloads—come in second and third.</p><h2>Training AI models can generate enormous carbon emissions</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Chart showing estimated carbon emissions from training of AI models from 2012 to 2025. With Grok 3 and Grok 4, the chart shows a tremendous increase in 2025." class="rm-shortcode" data-rm-shortcode-id="3e8a202e57d64ac69720c82cb8749d54" data-rm-shortcode-name="rebelmouse-image" id="08510" loading="lazy" src="https://spectrum.ieee.org/media-library/chart-showing-estimated-carbon-emissions-from-training-of-ai-models-from-2012-to-2025-with-grok-3-and-grok-4-the-chart-shows-a.jpg?id=65506058&width=980"/> </p><p><span>Stanford’s AI Index has called out the carbon emissions from AI training in prior years, and the issue continues to trend in a worrying direction.</span></p><p>The report estimates that training the latest frontier large language models, such as xAI’s Grok 4, can generate over 72,000 tons of carbon-equivalent emissions. That’s a huge increase from estimates in prior years. OpenAI’s GPT-4 was estimated at 5,184 tons, and Meta’s Llama 3.1 405B was estimated at 8,930 tons. </p><p><a href="https://hai.stanford.edu/people/ray-perrault" target="_blank">Ray Perrault</a>, co-director of the AI Index steering committee, says these figures are estimates. “These estimates should be interpreted with caution. In the case of Grok, they rely heavily on inferred inputs drawn from public reporting (e.g., <em>Forbes</em> articles), xAI statements, and other non-verifiable sources, introducing a degree of uncertainty,” says Perrault. On the other hand, Perrault noted that “Epoch AI independently estimates Grok 4’s emissions to be significantly higher at approximately 140,000 tons of CO₂.”</p><p>Emissions from AI inference also continue to increase, though results again vary by model. The report estimates that carbon emissions from models with the least efficient inference are over 10 times as high as those with the most efficient inference. <a data-linked-post="2671027978" href="https://spectrum.ieee.org/deepseek" target="_blank">DeepSeek</a>’s V3 models were estimated to consume around 23 watts when responding to a “medium-length” prompt, while Claude 4 Opus was estimated to consume about 5 watts.</p><h2>LLMs are rapidly defeating new benchmarks</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A chart shows AI index technical performance benchmarks compared to human performance for a variety of tasks from 2012 to 2025. Image classification surpassed human performance early, but only in the 2020s have models begun to near or surpass human baselines in a number of tasks." class="rm-shortcode" data-rm-shortcode-id="7a521cbb67087cb2bdd493f8982f990f" data-rm-shortcode-name="rebelmouse-image" id="8b302" loading="lazy" src="https://spectrum.ieee.org/media-library/a-chart-shows-ai-index-technical-performance-benchmarks-compared-to-human-performance-for-a-variety-of-tasks-from-2012-to-2025.jpg?id=65506063&width=980"/> </p><p><span>The capabilities of AI models have improved with incredible speed over the past decade, and as the graph above shows, progress seems to be accelerating. Multimodal LLMs, in particular, are conquering benchmarks nearly as quickly as they can be invented.</span></p><p><a data-linked-post="2669884140" href="https://spectrum.ieee.org/ai-agents" target="_blank">Agentic AI</a> has experienced the most extreme gains. The two steep lines at the right of the chart represent the <a href="https://os-world.github.io/" target="_blank">OSWorld benchmark</a>, which benchmarks autonomous computer use, and the <a href="https://openai.com/index/introducing-swe-bench-verified/" target="_blank">SWE-Bench Verified</a> software engineering benchmark, which benchmarks autonomous coding.</p><p>Models are also rapidly improving on <a href="https://agi.safe.ai/" target="_blank">Humanity’s Last Exam</a>. This benchmark includes questions contributed by subject-matter experts designed to represent the toughest problems in their fields. The 2025 Stanford AI Index reported the top-ranking model, OpenAI’s o1, correctly answered just 8.8 percent of questions. Since then, accuracy has increased to 38.3 percent—and even that number is a bit out of date, <a href="https://llm-stats.com/benchmarks/humanity's-last-exam" target="_blank">as the best-scoring models as of April 2026</a> (such as Anthropic’s Claude Opus 4.6 and Google’s Gemini 3.1 Pro) top 50 percent.</p><p>Still, Perrault cautioned that benchmarks may not always map to real-world results. “We generally lack measures of how well a system (or agent) needs to function in a particular setting,” says Perrault. “Knowing that a benchmark for legal reasoning has 75 percent accuracy tells us little about how well it would fit in a law practice’s activities.”</p><h2>AI research in medicine sees gains</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A bar graph shows increasing numbers of publications on AI for drug discovery from 2018 to 2025." class="rm-shortcode" data-rm-shortcode-id="b063563fc0db5ddc7f527ce88bca7399" data-rm-shortcode-name="rebelmouse-image" id="8120f" loading="lazy" src="https://spectrum.ieee.org/media-library/a-bar-graph-shows-increasing-numbers-of-publications-on-ai-for-drug-discovery-from-2018-to-2025.jpg?id=65506067&width=980"/> </p><p><span>Gains in AI benchmarks seem to be reflected in medicine, where AI adoption has increased at a rapid pace. Medical research shows particularly quick adoption. As the graph above shows, the number of publications on the topic of AI use for drug discovery has more than doubled over the past two years. There are 2.7 times as many publications on multimodal biomedical AI, which are used to examine medical images alongside text, as there were two years ago.</span></p><h2>LLMs still have trouble reading an analog clock</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A bar chart compares different LLMs taking on the task of reading an analog clock, ranging from only 8.9% to 50.60% accuracy." class="rm-shortcode" data-rm-shortcode-id="e106617ccc2cd8fb3fcf8cd9788f8543" data-rm-shortcode-name="rebelmouse-image" id="02fc0" loading="lazy" src="https://spectrum.ieee.org/media-library/a-bar-chart-compares-different-llms-taking-on-the-task-of-reading-an-analog-clock-ranging-from-only-8-9-to-50-60-accuracy.jpg?id=65506069&width=980"/> </p><p><span>While AI models have improved rapidly in some areas, they remain remarkably bad at some common tasks, like </span><a data-linked-post="2674259807" href="https://spectrum.ieee.org/large-language-models-reading-clocks" target="_blank">reading clocks</a><span> and understanding calendars. </span><a href="https://clockbench.ai/ClockBench.pdf" target="_blank">ClockBench</a><span>, which measures a multimodal LLM’s ability to read an analog clock, found that even the model best at this task, OpenAI’s GPT-5.4, had just 50-50 odds of getting it right.</span></p><p>Most models scored far worse. Anthropic’s Claude Opus 4.6 read the time correctly with just 8.9 percent accuracy. That’s surprising, because the model often scores well in other benchmarks. (As previously mentioned, Claude Opus 4.6 delivered top-notch scores in Humanity’s Last Exam.)</p><p>Of course, LLMs will rarely be asked to perform this task in real life, but Perrault says it represents a more general issue. “There is a research thread that shows that when systems are asked questions about combinations of language with other modalities (e.g., images, or audio, as in tone of voice), the language component carries a surprisingly large part of the burden, event to the extent of ignoring non-language information completely.” </p><h2>AI investment hit a new peak in 2025</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A bar chart showing global corporate investment in AI by investment activity from 2013 to 2025 highlighting a rise in 2021, followed by a dip in 2022-2024 and then a huge increase again in 2025." class="rm-shortcode" data-rm-shortcode-id="4801a130521bd0f38143f4882701c41c" data-rm-shortcode-name="rebelmouse-image" id="9dea4" loading="lazy" src="https://spectrum.ieee.org/media-library/a-bar-chart-showing-global-corporate-investment-in-ai-by-investment-activity-from-2013-to-2025-highlighting-a-rise-in-2021-foll.jpg?id=65506071&width=980"/> </p><p><span>The gains in AI model performance have gone hand-in-hand with investment in AI companies. According to data from AI analytics company </span><a href="https://www.quid.com/" target="_blank">Quid</a><span>, 2025 set a new record for AI investment with over US $581 billion spent.</span></p><p>That’s more than double the $253 billion spent in 2024 and speeds past the previous record of $360 billion, which was set in 2021. And unlike 2021, where investment was led by mergers and acquisitions, 2025’s record-setting result was led by private investment in AI companies.</p><p>Most of that money is flowing into the United States, where over $344 billion were invested in AI last year. </p><h2>Software engineers are all-in on AI</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A line graph shows the number of GitHub AI projects from 2011 to 2025 as a increase from about 0 to 5.58 million." class="rm-shortcode" data-rm-shortcode-id="6e2f54b974d8cb26eb62f6c48aa8e3bf" data-rm-shortcode-name="rebelmouse-image" id="db874" loading="lazy" src="https://spectrum.ieee.org/media-library/a-line-graph-shows-the-number-of-github-ai-projects-from-2011-to-2025-as-a-increase-from-about-0-to-5-58-million.jpg?id=65506074&width=980"/> </p><p><span>However, the story of AI adoption isn’t just about private money. There’s also a grassroots enthusiasm for AI on GitHub, where the number of AI-related projects has rocketed to 5.58 million projects through 2025. That’s a roughly fivefold increase since 2020 and a 23.7 percent increase from 2024.</span></p><p>This number doesn’t appear to represent a flood of AI-generated projects, either. The number of projects with at least 10 stars has increased at a similar rate, and the number of stars awarded to AI projects overall has increased at a similar rate. That suggests the projects are seeing human engagement. Perhaps this should be no surprise given the popularity of some projects. Open-source agentic AI software OpenClaw, for example, <a href="https://github.com/openclaw/openclaw" target="_blank">has received 352,000 stars</a>. </p><p>Critics may worry that the enthusiasm is in part driven by AI bots or agentic projects. Perrault acknowledged this and says that “probably the intensity of GitHub use is highly correlated with the intensity of AI use.” However, the majority of GitHub activity still appears to be conducted by humans, at least according to an activity-tracking website called <a href="https://insights.logicstar.ai">Agents in the Wild </a>(this website is not mentioned in Stanford’s report).</p><p>Enthusiasm is strong in computer science, too. The number of AI-related computer science publications has more than doubled over the past decade, from 102,000 to 258,000. More than 68 percent of these still originate in academia, with government and industry contributing about 11.5 and 12.5 percent, respectively as of 2024. The growth is led by publications in machine learning, computer vision, and generative AI.</p><h2>AI’s overall impact on employment remains unclear</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Two line charts show headcount trends for software developers and customer support agents by age from 2021 to 2025. Of note is a distinct decrease in headcount for early career workers." class="rm-shortcode" data-rm-shortcode-id="17e3f35d335584d1be06b2a46d40710b" data-rm-shortcode-name="rebelmouse-image" id="d66e9" loading="lazy" src="https://spectrum.ieee.org/media-library/two-line-charts-show-headcount-trends-for-software-developers-and-customer-support-agents-by-age-from-2021-to-2025-of-note-is-a.jpg?id=65506076&width=980"/> </p><p><span>The rise of generative AI goes hand-in-hand with employment worries, a phenomenon no doubt encouraged by the worrisome predictions of CEOs at the world’s largest AI companies. However, the data so far remains mixed.</span></p><p>Above you’ll find graphs that show the “normalized headcount” among varying age demographics in two professions thought to be at high risk of AI replacement: software developers and customer support agents. As in prior years, the trends show that entry-level jobs in these professions have been reduced, while mid-career and senior positions have held steady or increased. <br/><br/>However, these changes remain difficult to untangle from broader economic trends. The report notes that unemployment is rising across many occupations and that, contrary to expectations, unemployment among workers least exposed to AI has risen more than unemployment among workers most exposed to AI.</p><h2>Overall public perception of AI (slightly) improves</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Bar charts show responses to various opinion statements related to AI from 2022 to 2025. " class="rm-shortcode" data-rm-shortcode-id="a7fbb2f03e9c1954db8ae14e99f037a8" data-rm-shortcode-name="rebelmouse-image" id="33355" loading="lazy" src="https://spectrum.ieee.org/media-library/bar-charts-show-responses-to-various-opinion-statements-related-to-ai-from-2022-to-2025.jpg?id=65506086&width=980"/> </p><p><span>The report’s most surprising finding is, no doubt, the small but notable increase in optimism about AI over the past several years: 59 percent of respondents to a survey conducted by Ipsos said “the benefits outweigh the drawbacks,” up from 55 percent in 2024, and 68 percent of respondents said they have a “good understanding” of AI, a slight uptick from 67 percent in 2024.</span></p><p>Survey responses to similar questions suggest that the overall reception to AI is more positive than negative, though some negative feelings have also increased. For example, 52 percent of respondents said that products and services that use AI make them “nervous.”</p><p>Sentiment varies significantly by country. Countries in Southeast Asia, including China, Malaysia, Thailand, Indonesia, and Singapore, are trending more positive toward AI. However, the strongest positive year-over-year shifts were in Germany (12 percent), France (10 percent), and the Netherlands (10 percent). Colombia saw the most negative shift (-6 percent), a reversal of the trend from prior years.</p><h2>Trust in AI regulation varies significantly by country</h2><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A chart shows trust in government regulation of AI by country led by Singapore with 81% and with the United States at the bottom with 31%." class="rm-shortcode" data-rm-shortcode-id="f5c779dfa5e5b877b12121418d42b160" data-rm-shortcode-name="rebelmouse-image" id="6eb33" loading="lazy" src="https://spectrum.ieee.org/media-library/a-chart-shows-trust-in-government-regulation-of-ai-by-country-led-by-singapore-with-81-and-with-the-united-states-at-the-bottom.jpg?id=65506090&width=980"/> </p><p><span>While a growing number of people seem to feel that AI will have a positive impact, that shift is accompanied by deep distrust in some countries, particularly on the topic of government regulation.</span></p><p>Notably, the United States is at the bottom of the list even while it leads in AI investment. Only 31 percent of Ipsos survey respondents trusted the government to regulate AI. Many European countries showed low levels of trust, as did Japan. Countries in Asia and South America showed the greatest trust in their government’s ability to regulate AI.</p><p>The results from the United States and Colombia are intriguing. The U.S. is seeing deep distrust in AI regulation, yet most respondents think AI’s benefits will outweigh its drawbacks. Colombia, on the other hand, shows high trust in AI regulations yet worsening sentiment toward AI overall.</p><p>This feels like a microcosm of the AI narrative in 2025. Both the quality of results from AI models, and public perception on how AI will impact society, continue to vary, often by wide margins, depending on the task or question at hand. </p>]]></description><pubDate>Mon, 13 Apr 2026 13:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/state-of-ai-index-2026</guid><category>Ai-index</category><category>Artificial-intelligence</category><category>Stanford-university</category><dc:creator>Matthew S. Smith</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/squares-and-rectangles-on-graph-paper-form-the-letters-ai.jpg?id=65506010&amp;width=980"></media:content></item><item><title>GoZTASP: A Zero-Trust Platform for Governing Autonomous Systems at Mission Scale</title><link>https://content.knowledgehub.wiley.com/goztasp-a-zero-trust-platform-for-governing-autonomous-systems-at-mission-scale/</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/technology-innovation-institute-logo-with-stylized-tii-and-curved-line.png?id=65498963&width=980"/><br/><br/><p>ZTASP is a mission-scale assurance and governance platform designed for autonomous systems operating in real-world environments. It integrates heterogeneous systems—including drones, robots, sensors, and human operators—into a unified zero-trust architecture. Through Secure Runtime Assurance (SRTA) and Secure Spatio-Temporal Reasoning (SSTR), ZTASP continuously verifies system integrity, enforces safety constraints, and enables resilient operation even under degraded conditions.</p><p>ZTASP has progressed beyond conceptual design, with operational validation at Technology Readiness Level (TRL) 7 in mission critical environments. Core components, including Saluki secure flight controllers, have reached TRL8 and are deployed in customer systems. While initially developed for high-consequence mission environments, the same assurance challenges are increasingly present across domains such as healthcare, transportation, and critical infrastructure.</p><p><span><a href="https://content.knowledgehub.wiley.com/goztasp-a-zero-trust-platform-for-governing-autonomous-systems-at-mission-scale/" target="_blank">Download this free whitepaper now!</a></span></p>]]></description><pubDate>Thu, 09 Apr 2026 15:06:39 +0000</pubDate><guid>https://content.knowledgehub.wiley.com/goztasp-a-zero-trust-platform-for-governing-autonomous-systems-at-mission-scale/</guid><category>Autonomous-systems</category><category>Drones</category><category>Sensors</category><category>Transportation</category><category>Type-whitepaper</category><dc:creator>Technology Innovation Institute</dc:creator><media:content medium="image" type="image/png" url="https://assets.rbl.ms/65498963/origin.png"></media:content></item><item><title>AI Models Map the Colorado River’s Hard Choices</title><link>https://spectrum.ieee.org/colorado-river-water-shortage</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/overhead-view-of-horseshoe-bend-an-incised-meander-shaped-like-the-letter-u.jpg?id=65487612&width=1200&height=800&coordinates=156%2C0%2C156%2C0"/><br/><br/><p>The <a href="https://spectrum.ieee.org/tag/colorado-river" target="_blank">Colorado River</a> begins as snow. Every spring, the mountain snowpack of the Rockies melts into streams that feed into reservoirs that supply 40 million people across seven U.S. states. The system has worked, more or less, for a century. That century is over.</p><p>By some measures, 2026 is shaping up to be the worst year the river has seen since records began. Flows are down <a href="https://www.science.org/doi/10.1126/science.abj5498" rel="noopener noreferrer" target="_blank">20 percent from 2000 levels</a>. Lake Powell, the reservoir straddling Utah and Arizona, may drop below the threshold for generating hydropower <a href="https://coloradosun.com/2026/02/18/lake-powell-forecast-critical-lows-federal-study/" rel="noopener noreferrer" target="_blank">before the year is out</a>. The negotiations between the seven states over how to <a href="https://www.nytimes.com/2026/02/13/climate/colorado-river-cooperation-missed-deadline.html" rel="noopener noreferrer" target="_blank">share what’s left have collapsed twice</a>, and the U.S. federal government is threatening to impose its own plan.</p><p>While the states argue and the river shrinks, a growing set of machine learning tools is being deployed across the basin. Federal water managers are running millions of simulations to stress-test reservoir strategies against different possible futures. Researchers are forecasting streamflow months out using satellite data and deep learning. These technologies don’t promise to resolve the crisis, but they’re making the trade-offs visible. They’re showing, more precisely than ever before, what each decision will cost.</p><h2>Seeing Further Into the River’s Future</h2><p>Nobody manages more of the Colorado River’s daily operations than the <a href="https://www.usbr.gov/" rel="noopener noreferrer" target="_blank">U.S. Bureau of Reclamation</a>. If the federal government follows through on its threat to impose a water-sharing plan, it will be Reclamation doing the imposing, and making decisions about how much water flows from Lake Powell and Lake Mead, the two largest reservoirs in the country. </p><p>The agency is not new to sophisticated modeling. For years, Reclamation’s researchers have combined paleoclimate reconstructions, global circulation models, and scenario planning to predict the river’s future. Machine learning tools are adding to that toolkit, says <a href="https://www.linkedin.com/in/chris-frans-phd-pe-74491579/" target="_blank">Chris Frans</a>, Reclamation’s water-availability research coordinator, and they are already informing real operational decisions.</p><p>The clearest gains are in streamflow forecasting. Machine learning techniques—using data from satellites and weather stations well outside the basin—now outperform traditional methods across a range of conditions. Forecasts update every hour. In some areas, managers are getting five to seven days of advance warning on flood events, compared with three in the past, which gives them time to reduce the water in reservoirs before high inflows arrive.</p><p>The scale of scenario modeling has also expanded dramatically. A decade ago, running 100,000 individual simulations was a landmark study. Now, says <a href="https://www.colorado.edu/cadswes/alan-butler" rel="noopener noreferrer" target="_blank">Alan Butler</a>, who manages Reclamation’s research and modeling group for the lower Colorado Basin, millions of simulations feed the analytical tools used in the current guidelines. Those simulations map out how different operating strategies perform across widely varying futures—making the trade-offs between them harder to ignore.</p><h2>Dividing a Shrinking River</h2><p>Knowing how much water is coming is one problem. Deciding who gets it is another. At the center of that process is the <a href="https://coloradoriverscience.org/Colorado_River_Simulation_System_(CRSS)" rel="noopener noreferrer" target="_blank">Colorado River Simulation System</a> (CRSS), which models how water moves through the basin’s reservoirs, canals, and pipelines under more than a century of legal and regulatory constraints. This Reclamation model is an imperfect representation, but it has been the foundation of river negotiations for decades.</p><p>A tool called <a href="https://riverware.org/" rel="noopener noreferrer" target="_blank">RiverWare</a>, first developed in the early 1990s at the University of Colorado Boulder, lets states, cities, and tribes run their own scenarios through CRSS. Before RiverWare, these groups didn’t have confidence in Reclamation’s numbers, says <a href="https://www.colorado.edu/ceae/edith-zagona" rel="noopener noreferrer" target="_blank">Edith Zagona</a>, a Boulder professor who directs the <a href="https://www.colorado.edu/cadswes/" rel="noopener noreferrer" target="_blank">Center for Advanced Decision Support for Water and Environmental Systems</a>, the center that built it. “There was just this huge lack of trust.” The solution was letting stakeholders inspect the assumptions built into the RiverWare model—how much water was available, how it could be used, and under what rules. </p><p>Getting stakeholders to trust the model turned out to be the easier problem. The harder one is what to do when the model itself can’t predict a single probable future. That question drove Zagona toward a framework called decision-making under deep uncertainty, which trades prediction for stress-testing policies against thousands of possible futures.</p><p>The tool Zagona’s group developed with Reclamation and the consulting firm <a href="https://virgalabs.io/" rel="noopener noreferrer" target="_blank">Virga Labs</a> puts the framework into practice in a web-based tool, running CRSS across more than 8,000 possible future water-supply scenarios to show how different management strategies hold up against the full range of what climate change might bring. At its center is an evolutionary algorithm called Borg, which generates and iteratively refines those strategies, searching for plans that perform well across many scenarios. The result is a set of trade-offs, not a single answer. </p><p><a href="https://riverware.org/riverware/ugm/2025/PDFs/Users/10.kasprzyk-rwugm2025-borgRW-full.pdf" rel="noopener noreferrer" target="_blank">Borg-RiverWare</a> has already shaped the ongoing negotiations over the river’s next operating rules, generating the scenarios and data that Reclamation used in its modeling tools. Those tools give stakeholders a common analytical foundation for negotiations. Now Zagona’s center is pushing the approach further. A system in development would let negotiating parties test competing proposals on the fly, showing how one side’s policy choices would ripple through the system and identifying areas of potential compromise during the negotiation itself.</p><h2>New Tools for Forecasting the Colorado</h2><p>Reclamation and Zagona’s center aren’t the only ones trying to see further into the river’s future. At Metropolitan State University of Denver, a team led by <a href="https://red.msudenver.edu/expert/mohammad-valipour/" rel="noopener noreferrer" target="_blank">Mohammad Valipour</a> has been building a forecasting system that uses deep learning to issue drought warnings across seven rivers in Colorado, from seven days to six months out. In a region where ground gauges are sparse and mountains make installation difficult, the team found that NASA satellite data outperformed in-field measurements. The goal, Valipour says, is a statewide drought alarm system that gives farmers and water managers more time to respond.</p><p>At Utah State University, <a href="https://engineering.usu.edu/cs/directory/faculty/boubrahimi-filali-soukaina" rel="noopener noreferrer" target="_blank">Soukaina Filali Boubrahimi</a> is attacking a different problem: how conditions at one point in the river ripple downstream weeks later. Using a graph neural network that treats each monitoring station as a node, her team built a map of the river’s interdependencies across one of the most contested water systems in the world. She says the approach could extend to other overtaxed basins.</p><p>“If you can figure out the Colorado River,” she says, “anyone else dealing with a stressed river system is going to be interested in what you learned.”</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Line graph of snow water equivalent projection in the upper Colorado region. As of March 26th 2026, most maximum projections through June still fall short of the median." class="rm-shortcode" data-rm-shortcode-id="51c77aad5a41503f03b2d3ade20069a1" data-rm-shortcode-name="rebelmouse-image" id="07b94" loading="lazy" src="https://spectrum.ieee.org/media-library/line-graph-of-snow-water-equivalent-projection-in-the-upper-colorado-region-as-of-march-26th-2026-most-maximum-projections-thr.jpg?id=65487654&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Snowpack in the upper Colorado River basin is far below normal. As of late March 2026, measurements across 130 sites were about 35 percent of the median, with projections showing continued shortfalls.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit..."><a href="https://nwcc-apps.sc.egov.usda.gov/awdb/basin-plots/Proj/WTEQ/assocHUC2/14_Upper_Colorado_Region.html?hucFilter=14" target="_blank">USDA Natural Resources Conservation Service (NRCS)</a></small></p><h2>What the Models Can’t See</h2><p>Across the basin, researchers and water managers are running into the same wall. The models learn from historical data, but that data describes a river that no longer exists. Valipour found that feeding his models only the last decade outperformed using longer records. Filali Boubrahimi’s model struggles most in drought conditions, precisely when predictions matter most, because recent prolonged droughts don’t resemble the historical training data. One workaround is to train models on data from basins that have already experienced what the Colorado hasn’t yet.</p><p>Even so, better forecasts do not resolve the central problem. While the tools can show you what a drier future looks like across a thousand possible scenarios, they can’t tell you who should bear the cost of it. The cuts coming to the basin are going to be enormous, says <a href="https://www.colorado.edu/center/gwc/brad-udall" target="_blank">Brad Udall</a>, a water and climate research scientist at Colorado State University’s <a href="https://watercenter.colostate.edu/" target="_blank">Colorado Water Center</a>, and they will fall mostly on <a href="https://spectrum.ieee.org/tag/agriculture" target="_blank">agriculture</a>. They may fundamentally reshape communities that have built their economies around water for generations. “AI has no business being in the realm of replacing human values and human judgments,” he says.</p><p>The tools, by most measures, are doing exactly what they were built to do: The negotiating parties understand what is coming, and they are not disputing the projections. Zagona, who has worked on the Colorado River for 45 years, sees reasons for optimism. “The tools are bringing people to the table,” she says. “They’re at the table arguing. But at least they’re at the table.”</p>]]></description><pubDate>Wed, 08 Apr 2026 14:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/colorado-river-water-shortage</guid><category>Colorado-river</category><category>Drought</category><category>Environmental-policy</category><category>Climate-change</category><category>Simulations</category><category>Evolutionary-algorithm</category><dc:creator>Jackie Snow</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/overhead-view-of-horseshoe-bend-an-incised-meander-shaped-like-the-letter-u.jpg?id=65487612&amp;width=980"></media:content></item><item><title>Decentralized Training Can Help Solve AI’s Energy Woes</title><link>https://spectrum.ieee.org/decentralized-ai-training-2676670858</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/illustration-of-several-data-servers-interconnected-across-long-distances.jpg?id=65477795&width=1200&height=800&coordinates=156%2C0%2C156%2C0"/><br/><br/><p> <a href="https://spectrum.ieee.org/topic/artificial-intelligence/" target="_self">Artificial intelligence</a> harbors an enormous <a href="https://spectrum.ieee.org/topic/energy/" target="_self">energy</a> appetite. Such constant cravings are evident in the <a href="https://spectrum.ieee.org/ai-index-2025" target="_self">hefty carbon footprint</a> of the <a href="https://spectrum.ieee.org/tag/data-centers" target="_self">data centers</a> behind the AI boom and the steady increase over time of <a href="https://spectrum.ieee.org/tag/carbon-emissions" target="_self">carbon emissions</a> from training frontier <a href="https://spectrum.ieee.org/tag/ai-models" target="_self">AI models</a>.</p><p>No wonder big tech companies are warming up to <a href="https://spectrum.ieee.org/tag/nuclear-energy" target="_self">nuclear energy</a>, envisioning a future fueled by reliable, carbon-free sources. But while <a href="https://spectrum.ieee.org/nuclear-powered-data-center" target="_self">nuclear-powered data centers</a> might still be years away, some in the research and industry spheres are taking action right now to curb AI’s growing energy demands. They’re tackling training as one of the most energy-intensive phases in a model’s life cycle, focusing their efforts on decentralization.</p><p>Decentralization allocates model training across a network of independent nodes rather than relying on one platform or provider. It allows compute to go where the energy is—be it a dormant server sitting in a research lab or a computer in a <a href="https://spectrum.ieee.org/tag/solar-power" target="_self">solar-powered</a> home. Instead of constructing more data centers that require <a href="https://spectrum.ieee.org/tag/power-grid" target="_self">electric grids</a> to scale up their infrastructure and capacity, decentralization harnesses energy from existing sources, avoiding adding more power into the mix.</p><h2>Hardware in harmony</h2><p>Training AI models is a huge data center sport, synchronized across clusters of closely connected <a href="https://spectrum.ieee.org/tag/gpus" target="_self">GPUs</a>. But as <a href="https://spectrum.ieee.org/mlperf-trends" target="_self">hardware improvements struggle to keep up</a> with the swift rise in the size of <a href="https://spectrum.ieee.org/tag/large-language-models" target="_self">large language models</a>, even massive single data centers are no longer cutting it.</p><p>Tech firms are turning to the pooled power of multiple data centers—no matter their location. <a href="https://spectrum.ieee.org/tag/nvidia" target="_self">Nvidia</a>, for instance, launched the <a href="https://developer.nvidia.com/blog/how-to-connect-distributed-data-centers-into-large-ai-factories-with-scale-across-networking/" target="_blank">Spectrum-XGS Ethernet for scale-across networking</a>, which “can deliver the performance needed for large-scale single job AI training and inference across geographically separated data centers.” Similarly, <a href="https://spectrum.ieee.org/tag/cisco" target="_self">Cisco</a> introduced its <a href="https://blogs.cisco.com/sp/the-new-benchmark-for-distributed-ai-networking" target="_blank">8223 router</a> designed to “connect geographically dispersed AI clusters.”</p><p>Other companies are harvesting idle compute in <a href="https://spectrum.ieee.org/tag/servers" target="_self">servers</a>, sparking the emergence of a <a href="https://spectrum.ieee.org/gpu-as-a-service" target="_self">GPU-as-a-Service</a> business model. Take <a href="https://akash.network/" rel="noopener noreferrer" target="_blank">Akash Network</a>, a peer-to-peer <a href="https://spectrum.ieee.org/tag/cloud-computing" target="_self">cloud computing</a> marketplace that bills itself as the “Airbnb for data centers.” Those with unused or underused GPUs in offices and smaller data centers register as providers, while those in need of computing power are considered as tenants who can choose among providers and rent their GPUs.</p><p>“If you look at [AI] training today, it’s very dependent on the latest and greatest GPUs,” says Akash cofounder and CEO <a href="https://www.linkedin.com/in/gosuri" rel="noopener noreferrer" target="_blank">Greg Osuri</a>. “The world is transitioning, fortunately, from only relying on large, high-density GPUs to now considering smaller GPUs.”</p><h2>Software in sync</h2><p>In addition to orchestrating the <a href="https://spectrum.ieee.org/tag/hardware" target="_self">hardware</a>, decentralized AI training also requires algorithmic changes on the <a href="https://spectrum.ieee.org/tag/software" target="_self">software</a> side. This is where <a href="https://cloud.google.com/discover/what-is-federated-learning" rel="noopener noreferrer" target="_blank">federated learning</a>, a form of distributed <a href="https://spectrum.ieee.org/tag/machine-learning" target="_self">machine learning</a>, comes in.</p><p>It starts with an initial version of a global AI model housed in a trusted entity such as a central server. The server distributes the model to participating organizations, which train it locally on their data and share only the model weights with the trusted entity, explains <a href="https://www.csail.mit.edu/person/lalana-kagal" rel="noopener noreferrer" target="_blank">Lalana Kagal</a>, a principal research scientist at <a href="https://www.csail.mit.edu/" rel="noopener noreferrer" target="_blank">MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL)</a> who leads the <a href="https://www.csail.mit.edu/research/decentralized-information-group-dig" rel="noopener noreferrer" target="_blank">Decentralized Information Group</a>. The trusted entity then aggregates the weights, often by averaging them, integrates them into the global model, and sends the updated model back to the participants. This collaborative training cycle repeats until the model is considered fully trained.</p><p>But there are drawbacks to distributing both data and computation. The constant back-and-forth exchanges of model weights, for instance, result in high communication costs. Fault tolerance is another issue.</p><p>“A big thing about AI is that every training step is not fault-tolerant,” Osuri says. “That means if one node goes down, you have to restore the whole batch again.”</p><p>To overcome these hurdles, researchers at <a href="https://deepmind.google/" rel="noopener noreferrer" target="_blank">Google DeepMind</a> developed <a href="https://arxiv.org/abs/2311.08105" rel="noopener noreferrer" target="_blank">DiLoCo</a>, a distributed low-communication optimization <a href="https://spectrum.ieee.org/tag/algorithms" target="_self">algorithm</a>. DiLoCo forms what <a href="https://spectrum.ieee.org/tag/google-deepmind" target="_self">Google DeepMind</a> research scientist <a href="https://arthurdouillard.com/" rel="noopener noreferrer" target="_blank">Arthur Douillard</a> calls “islands of compute,” where each island consists of a group of <a href="https://spectrum.ieee.org/tag/chips" target="_self">chips</a>. Every island holds a different chip type, but chips within an island must be of the same type. Islands are decoupled from each other, and synchronizing knowledge between them happens once in a while. This decoupling means islands can perform training steps independently without communicating as often, and chips can fail without having to interrupt the remaining healthy chips. However, the team’s experiments found diminishing performance after eight islands.</p><p>An improved version, dubbed <a href="https://arxiv.org/abs/2501.18512" rel="noopener noreferrer" target="_blank">Streaming DiLoCo</a>, further reduces the bandwidth requirement by synchronizing knowledge “in a streaming fashion across several steps and without stopping for communicating,” says Douillard. The mechanism is akin to watching a video even if it hasn’t been fully downloaded yet. “In Streaming DiLoCo, as you do computational work, the knowledge is being synchronized gradually in the background,” he adds.</p><p>AI development platform <a href="https://www.primeintellect.ai/" rel="noopener noreferrer" target="_blank">Prime Intellect</a> implemented a variant of the DiLoCo algorithm as a vital component of its 10-billion-parameter <a href="https://www.primeintellect.ai/blog/intellect-1-release" rel="noopener noreferrer" target="_blank">INTELLECT-1</a> model trained across five countries spanning three continents. Upping the ante, <a href="https://0g.ai/" rel="noopener noreferrer" target="_blank">0G Labs</a>, makers of a decentralized AI <a href="https://spectrum.ieee.org/tag/operating-system" target="_self">operating system</a>, <a href="https://0g.ai/blog/worlds-first-distributed-100b-parameter-ai" rel="noopener noreferrer" target="_blank">adapted DiLoCo to train a 107-billion-parameter foundation model</a> under a network of segregated clusters with limited bandwidth. Meanwhile, popular <a href="https://spectrum.ieee.org/tag/open-source" target="_self">open-source</a> <a href="https://spectrum.ieee.org/tag/deep-learning" target="_self">deep learning</a> framework <a href="https://pytorch.org/projects/pytorch/" rel="noopener noreferrer" target="_blank">PyTorch</a> included DiLoCo in its <a href="https://meta-pytorch.org/torchft/" rel="noopener noreferrer" target="_blank">repository of fault-tolerance techniques</a>.</p><p>“A lot of engineering has been done by the community to take our DiLoCo paper and integrate it in a system learning over consumer-grade internet,” Douillard says. “I’m very excited to see my research being useful.”</p><h2>A more energy-efficient way to train AI</h2><p>With hardware and software enhancements in place, decentralized AI training is primed to help solve AI’s energy problem. This approach offers the option of training models “in a cheaper, more resource-efficient, more energy-efficient way,” says MIT CSAIL’s Kagal.</p><p>And while Douillard admits that “training methods like DiLoCo are arguably more complex, they provide an interesting trade-off of system efficiency.” For instance, you can now use data centers across far apart locations without needing to build ultrafast bandwidth in between. Douillard adds that fault tolerance is baked in because “the blast radius of a chip failing is limited to its island of compute.”</p><p>Even better, companies can take advantage of existing underutilized processing capacity rather than continuously building new energy-hungry data centers. Betting big on such an opportunity, Akash created its <a href="https://www.youtube.com/watch?v=zAj41xSNPeI" rel="noopener noreferrer" target="_blank">Starcluster program</a>. One of the program’s aims involves tapping into solar-powered homes and employing the desktops and laptops within them to train AI models. “We want to convert your home into a fully functional data center,” Osuri says.</p><p>Osuri acknowledges that participating in Starcluster will not be trivial. Beyond solar panels and devices equipped with consumer-grade GPUs, participants would also need to invest in <a href="https://spectrum.ieee.org/tag/batteries" target="_self">batteries</a> for backup power and redundant internet to prevent downtime. The Starcluster program is figuring out ways to package all these aspects together and make it easier for homeowners, including collaborating with industry partners to subsidize battery costs.</p><p>Back-end work is already underway to enable <a href="https://akash.network/roadmap/aep-60/" rel="noopener noreferrer" target="_blank">homes to participate as providers in the Akash Network</a>, and the team hopes to reach its target by 2027. The Starcluster program also envisions expanding into other solar-powered locations, such as schools and local community sites.</p><p>Decentralized AI training holds much promise to steer AI toward a more environmentally sustainable future. For Osuri, such potential lies in moving AI “to where the energy is instead of moving the energy to where AI is.”</p>]]></description><pubDate>Tue, 07 Apr 2026 14:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/decentralized-ai-training-2676670858</guid><category>Training</category><category>Ai-energy</category><category>Data-center</category><category>Large-language-models</category><dc:creator>Rina Diane Caballar</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/illustration-of-several-data-servers-interconnected-across-long-distances.jpg?id=65477795&amp;width=980"></media:content></item><item><title>Why AI Systems Fail Quietly</title><link>https://spectrum.ieee.org/ai-reliability</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-series-of-135-green-dots-slowly-transition-from-bright-green-to-black.png?id=65461614&width=1200&height=800&coordinates=73%2C0%2C74%2C0"/><br/><br/><p>In late-stage testing of a distributed AI platform, engineers sometimes encounter a perplexing situation: Every monitoring dashboard reads “healthy,” yet users report that the system’s decisions are slowly becoming wrong.</p><p>Engineers are trained to recognize <a href="https://spectrum.ieee.org/amp/it-management-software-failures-2674305315" target="_blank">failure</a> in familiar ways: a service crashes, a sensor stops responding, a constraint violation triggers a shutdown. Something breaks, and the system tells you. But a growing class of software failures looks very different. The system keeps running, logs appear normal, and monitoring dashboards stay green. Yet the system’s behavior quietly drifts away from what it was designed to do.</p><p>This pattern is becoming more common as autonomy spreads across software systems. Quiet failure is emerging as one of the defining engineering challenges of autonomous systems because correctness now depends on coordination, timing, and feedback across entire systems.</p><h2>When Systems Fail Without Breaking</h2><p>Consider a hypothetical enterprise AI assistant designed to summarize regulatory updates for financial analysts. The system retrieves documents from internal repositories, synthesizes them using a language model, and distributes summaries across internal channels.</p><p>Technically, everything works. The system retrieves valid documents, generates coherent summaries, and delivers them without issue.</p><p>But over time, something slips. Maybe an updated document repository isn’t added to the retrieval pipeline. The assistant keeps producing summaries that are coherent and internally consistent, but they’re increasingly based on obsolete information. Nothing crashes, no alerts fire, every component behaves as designed. The problem is that the overall result is wrong.</p><p>From the outside, the system looks operational. From the perspective of the organization relying on it, the system is quietly failing.</p><h2>The Limits of Traditional Observability</h2><p>One reason quiet failures are difficult to detect is that traditional systems measure the wrong signals. Operational dashboards track uptime, latency, and error rates, the core elements of modern <a href="https://www.ibm.com/think/topics/observability" target="_blank">observability</a>. These metrics are well-suited for transactional applications where requests are processed independently, and correctness can often be verified immediately.</p><p>Autonomous systems behave differently. Many AI-driven systems operate through continuous reasoning loops, where each decision influences subsequent actions. Correctness emerges not from a single computation but from sequences of interactions across components and over time. A retrieval system may return contextually inappropriate and technically valid information. A <a href="https://spectrum.ieee.org/ai-agent-benchmarks" target="_blank">planning agent</a> may generate steps that are locally reasonable but globally unsafe. A distributed decision system may execute correct actions in the wrong order.</p><p>None of these conditions necessarily produces errors. From the perspective of conventional observability, the system appears healthy. From the perspective of its intended purpose, it may already be failing.</p><h2>Why Autonomy Changes Failure</h2><p>The deeper issue is architectural. Traditional software systems were built around discrete operations: a request arrives, the system processes it, and the result is returned. Control is episodic and externally initiated by a user, scheduler, or external trigger.</p><p>Autonomous systems change that structure. Instead of responding to individual requests, they observe, reason, and act continuously. AI agents maintain context across interactions. Infrastructure systems adjust resources in real time. Automated workflows trigger additional actions without human input.</p><p>In these systems, correctness depends less on whether any single component works and more on coordination across time.</p><p>Distributed-systems engineers have long wrestled with issues of coordination. But this is coordination of a new kind. It’s no longer about things like keeping data consistent across services. It’s about ensuring that a stream of decisions—made by models, reasoning engines, planning algorithms, and tools, all operating with partial context—adds up to the right outcome.</p><p>A modern AI system may evaluate thousands of signals, generate candidate actions, and execute them across a distributed infrastructure. Each action changes the environment in which the next decision is made. Under these conditions, small <a href="https://spectrum.ieee.org/ai-mistakes-schneier" target="_blank">mistakes</a> can compound. A step that is locally reasonable can still push the system further off course.</p><p>Engineers are beginning to confront what might be called behavioral reliability: whether an autonomous system’s actions remain aligned with its intended purpose over time.</p><h2>The Missing Layer: Behavioral Control</h2><p>When organizations encounter quiet failures, the initial instinct is to improve monitoring: deeper logs, better tracing, more analytics. Observability is essential, but it only shows that the behavior has already diverged—it doesn’t correct it.</p><p>Quiet failures require something different: the ability to shape system behavior while it is still unfolding. In other words, autonomous systems increasingly need control architectures, not just monitoring.</p><p>Engineers in industrial domains have long relied on <a href="https://en.wikipedia.org/wiki/Supervisory_control" target="_blank">supervisory control systems</a>. These are software layers that continuously evaluate a system’s status and intervene when behavior drifts outside safe bounds. Aircraft flight-control systems, power-grid operations, and large manufacturing plants all rely on such supervisory loops. Software systems historically avoided them because most applications didn’t need them. Autonomous systems increasingly do.</p><p>Behavioral monitoring in AI systems focuses on whether actions remain aligned with intended purpose, not just whether components are functioning. Instead of relying only on metrics such as latency or error rates, engineers look for signs of behavior drift: <a href="https://en.wikipedia.org/wiki/Concept_drift" target="_blank">shifts in outputs</a>, inconsistent handling of similar inputs, or changes in how multistep tasks are carried out. An AI assistant that begins citing outdated sources, or an automated system that takes corrective actions more often than expected, may signal that the system is no longer using the right information to make decisions. In practice, this means tracking outcomes and patterns of behavior over time.</p><p>Supervisory control builds on these signals by intervening while the system is running. A supervisory layer checks whether ongoing actions remain within acceptable bounds and can respond by delaying or blocking actions, limiting the system to safer operating modes, or routing decisions for review. In more advanced setups, it can adjust behavior in real time—for example, by restricting data access, tightening constraints on outputs, or requiring extra confirmation for high-impact actions.</p><p>Together, these approaches turn reliability into an active process. Systems don’t just run, they are continuously checked and steered. Quiet failures may still occur, but they can be detected earlier and corrected while the system is operating.</p><h2>A Shift in Engineering Thinking</h2><p>Preventing quiet failures requires a shift in how engineers think about reliability: from ensuring components work correctly to ensuring system behavior stays aligned over time. Rather than assuming that correct behavior will emerge automatically from component design, engineers must increasingly treat behavior as something that needs active supervision.</p><p>As AI systems become more autonomous, this shift will likely spread across many domains of computing, including cloud infrastructure, robotics, and large-scale decision systems. The hardest engineering challenge may no longer be building systems that work, but ensuring that they continue to do the right thing over time.</p>]]></description><pubDate>Tue, 07 Apr 2026 13:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/ai-reliability</guid><category>Software-failure</category><category>Software-reliability</category><category>Software-engineering</category><category>Cloud-computing</category><category>Autonomous-systems</category><dc:creator>Varun Raj</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/a-series-of-135-green-dots-slowly-transition-from-bright-green-to-black.png?id=65461614&amp;width=980"></media:content></item><item><title>AI Is Insatiable</title><link>https://spectrum.ieee.org/high-bandwidth-memory-shortage</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/robot-hand-catching-falling-computer-chips-from-an-open-snack-bag-in-pop-art-style.png?id=65425799&width=1200&height=800&coordinates=0%2C180%2C0%2C181"/><br/><br/><p>While browsing our website a few weeks ago, I stumbled upon “<a href="https://spectrum.ieee.org/dram-shortage" target="_self">How and When the Memory Chip Shortage Will End</a>” by Senior Editor Samuel K. Moore. His analysis focuses on the current DRAM shortage caused by AI hyperscalers’ ravenous appetite for memory, a major constraint on the speed at which large language models run. Moore provides a clear explanation of the shortage, particularly for high bandwidth memory (HBM).</p><p>As we and the rest of the tech media have documented, AI is a resource hog. AI <a href="https://spectrum.ieee.org/data-center-sustainability-metrics" target="_self">electricity consumption</a> could account for up to 12 percent of all U.S. power by 2028. <a href="https://spectrum.ieee.org/ai-energy-use" target="_self">Generative AI queries</a> consumed 15 terawatt-hours in 2025 and are projected to consume 347 TWh by 2030. <a href="https://spectrum.ieee.org/data-centers-pollution" target="_self">Water consumption for cooling AI data centers</a> is predicted to double or even quadruple by 2028 compared to 2023.</p><p>But Moore’s reporting shines a light on an obscure corner of the AI boom. <a href="https://spectrum.ieee.org/processing-in-dram-accelerates-ai" target="_self">HBM</a> is a particular type of memory product tailor-made to serve AI processors. Makers of those processors, notably Nvidia and AMD, are demanding more and more memory for each of their chips, driven by the needs and wants of firms like Google, Microsoft, OpenAI, and Anthropic, which are underwriting an unprecedented buildout of data centers. And some of these facilities are colossal: You can read about the engineering challenges of building Meta’s mind-boggling 5-gigawatt Hyperion site in Louisiana, in “<a href="https://spectrum.ieee.org/5gw-data-center" target="_blank">What Will It Take to Build the World’s Largest Data Center?</a>”</p><p>We realized that Moore’s HBM story was both important and unique, and so we decided to include it in this issue, with some updates since the original published on 10 February. We paired it with a recent story by Contributing Editor Matthew S. Smith exploring how the memory-chip shortage is driving up the price of low-cost computers like the <a href="https://www.raspberrypi.com/" rel="noopener noreferrer" target="_blank">Raspberry Pi</a>. The result is “<a href="https://spectrum.ieee.org/dram-shortage" target="_blank">AI Is a Memory Hog</a>.”</p><p>The big question now is, When will the shortage end? Price pressure caused by AI hyperscaler demand on all kinds of consumer electronics is being masked by stubborn inflation combined with a perpetually shifting tariff regime, at least here in the United States. So I asked Moore what indicators he’s looking for that would signal an easing of the memory shortage.</p><p>“On the supply side, I’d say that if any of the big three HBM companies—<a href="https://www.micron.com/" rel="noopener noreferrer" target="_blank">Micron</a>, <a href="https://semiconductor.samsung.com/dram/" rel="noopener noreferrer" target="_blank">Samsung</a>, and <a href="https://www.skhynix.com/" rel="noopener noreferrer" target="_blank">SK Hynix</a>—say that they are adjusting the schedule of the arrival of new production, that’d be an important signal,” Moore told me. “On the demand side, it will be interesting to see how tech companies adapt up and down the supply chain. Data centers might steer toward hardware that sacrifices some performance for less memory. Startups developing all sorts of products might pivot toward creative redesigns that use less memory. Constraints like shortages can lead to interesting technology solutions, so I’m looking forward to covering those.”</p><p><span>To be sure you don’t miss any of Moore’s analysis of this topic and to stay current on the entire spectrum of technology development, <a href="https://spectrum.ieee.org/newsletters/" target="_blank">sign up for our weekly newsletter, Tech Alert.</a></span></p>]]></description><pubDate>Mon, 06 Apr 2026 14:22:58 +0000</pubDate><guid>https://spectrum.ieee.org/high-bandwidth-memory-shortage</guid><category>Semiconductors</category><category>Dram</category><category>Memory</category><category>Chips</category><category>Ai</category><category>Data-centers</category><dc:creator>Harry Goldstein</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/robot-hand-catching-falling-computer-chips-from-an-open-snack-bag-in-pop-art-style.png?id=65425799&amp;width=980"></media:content></item><item><title>The AI Data Centers That Fit on a Truck</title><link>https://spectrum.ieee.org/modular-data-center</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/overhead-view-of-two-data-center-pods-each-measuring-55-feet-long-by-12-5-feet-wide.jpg?id=65417343&width=1200&height=800&coordinates=0%2C83%2C0%2C84"/><br/><br/><p>A <a data-linked-post="2676577917" href="https://spectrum.ieee.org/5gw-data-center" target="_blank">traditional</a> data center protects the expensive hardware inside it with a “shell” constructed from steel and concrete. Constructing a data center’s shell is inexpensive compared to the cost of the hardware and infrastructure inside it, but it’s not trivial. It takes time for engineers to consider potential sites, apply for permits, and coordinate with construction contractors.</p><p>That’s a problem for those looking to quickly deploy AI hardware, which has led companies like <a href="https://duosedge.ai/home" target="_blank">Duos Edge AI</a> and <a href="https://www.lgcns.com/en" target="_blank">LG CNS</a> to respond with a more modular approach. They use pre-fabricated, self-contained boxes that can be deployed in months instead of years. The boxes can operate alone or in tandem with others, providing the option to add more if required.</p><p>“I just came back from Nvidia’s GTC, and a lot of [companies] are sitting on their deployment because their data centers aren’t ready, or they can’t find the space,” said <a href="https://www.linkedin.com/in/doug-recker/" rel="noopener noreferrer" target="_blank">Doug Recker</a>, CEO of Duos Edge AI. “We see the demand there, and we can deploy faster.” </p><h2>GPUs shipped straight to you</h2><p>Duos Edge AI’s modular compute pods are 55 feet long and 12.5 feet wide. Though they look similar to a shipping container, they’re actually a bit larger and designed primarily for transportation by truck. Each compute pod contains racks of GPUs much like those used in other data centers. Duos recently <a href="https://ir.duostechnologies.com/news-events/press-releases/detail/830/duos-technologies-group-executes-definitive-agreement-with" target="_blank">entered</a> a deal with AI infrastructure company Hydra Host to deploy four pods with 576 GPUs per pod. That’s a total of 2,304 GPUs, with the option to later double the deployment to 4,608 GPUs. </p><p>Modular data centers aren’t new for Duos; the company previously deployed edge data centers for rural customers, <a href="https://spectrum.ieee.org/rural-data-centers" target="_self">such as the Amarillo, Texas, school district</a>. However, the pods for the Hydra Host deployment will be upgraded to handle more intense AI workloads. They’ll contain more racks, draw more power, and use liquid cooling to keep the GPUs running efficiently. <br/><br/>Across the Pacific, Korean technology giant LG is taking a similar approach. The company’s CNS subsidiary, which provides IT infrastructure and services, <a href="https://www.koreatimes.co.kr/business/tech-science/20260305/lg-cns-unveils-container-based-ai-box-for-rapid-ai-data-center-expansion">has announced the AI Modular Data Center, which</a>, like the Duos unit, contains racks of GPUs and supporting hardware in a pre-fabricated enclosure.</p><p>Also like Duos’ deployment, LG’s AI Modular Data Center contains 576 Nvidia GPUs with the option to scale up in the future. “We are currently developing an expanded version that can support more than 4,600 GPUs within a single unit, with a service launch planned within this year,” said <a href="https://www.linkedin.com/in/heonhyeock-cho-29427b147/?originalSubdomain=kr" rel="noopener noreferrer" target="_blank">Heon Hyeock Cho</a>, vice president and head of the data center business unit at LG CNS. LG’s first Modular Data Center will roll out in the South Korean port city of Busan, where it could deploy up to 50 units.</p><p>LG and Duos are not alone. <a href="https://www.hpe.com/us/en/services/ai-mod-pod.html" rel="noopener noreferrer" target="_blank">Hewlett Packard Enterprise,</a> <a href="https://www.vertiv.com/en-emea/solutions/vertiv-modular-solutions/?utm_source=press-release&utm_medium=public-relations&utm_campaign=hpc-ai&utm_content=en-coolchip" rel="noopener noreferrer" target="_blank">Vertiv</a>, and <a href="https://www.se.com/ww/en/work/solutions/data-centers-and-networks/modular-data-center/" rel="noopener noreferrer" target="_blank">Schneider Electric</a> now have modular data centers available or in development. A <a href="https://www.grandviewresearch.com/industry-analysis/modular-data-center-market-report" target="_blank">report</a> from market research firm <a href="https://www.grandviewresearch.com/" target="_blank">Grand View Research</a> estimates that the market for modular data centers could more than double by 2030.</p><h2>On the grid, but under the radar</h2><p>A modular data center site is quite different from a traditional data center because there’s no need to construct a large steel-and-concrete shell. Instead, the site can be made ready by pouring a concrete pad. The pre-fabricated modules are delivered by truck, placed on the pad where desired, and then networked on-site.<br/><br/>Duos’ deployments, for instance, include power modules placed alongside the compute pods, and the pods are networked together with redundant fiber connections that allow the pods to operate in unison. Recker compared it to lining up school buses in a parking lot. “Everything is built off-site at a factory, and we can put it together like a jigsaw puzzle,” he said.</p><p>That simplicity is the point. Both Duos and LG CNS expect a modular data center can be deployed in about six months, compared to the roughly two or three years a conventional data center requires. Recker said that, for Duos, the turnaround is so quick that building the pre-fabricated unit isn’t always the constraint. While it’s possible to construct a pre-fabricated unit in 60 or 90 days, site preparation extends the timeline “because you can’t get the permits that fast.”</p><p>Modular data centers may also provide good value. Recker said a 5-megawatt modular deployment can be built for about $25 million, and that Duos’ cost per megawatt is roughly half what larger facilities charge. For Duos, savings are possible in part because its modular data centers can target smaller deployments where the permitting is less complex. Smaller, modular deployments also meet less resistance from local governments, which are increasingly skeptical about data center construction. </p><p>While Duos targets smaller deployments, LG hopes to go big. Its planned Busan campus of 50 AI Modular Data Centers suggests an ambition to achieve deployments that rival the capacity of conventional facilities. A site with 50 units would bring the total number of GPUs to over 28,000. Here, the benefits of a modular approach could stem mostly from scalability, as a modular data center could start small and grow as required.</p><p>“By adopting a modular approach, the AI Modular Data Center can be incrementally expanded through the combination of dozens of AI Boxes,” Cho said. “It’s enabling the construction of even hyperscale-level AI data centers.”</p>]]></description><pubDate>Mon, 30 Mar 2026 14:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/modular-data-center</guid><category>Data-center</category><category>Networking</category><category>Liquid-cooling</category><category>Ai</category><dc:creator>Matthew S. Smith</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/overhead-view-of-two-data-center-pods-each-measuring-55-feet-long-by-12-5-feet-wide.jpg?id=65417343&amp;width=980"></media:content></item><item><title>Why Are Large Language Models So Terrible at Video Games?</title><link>https://spectrum.ieee.org/ai-video-games-llms-togelius</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-middle-aged-man-smiling-while-holding-a-video-game-controller-behind-him-is-a-bookshelf-filled-with-countless-cooperative-boa.jpg?id=65413486&width=1200&height=800&coordinates=0%2C208%2C0%2C209"/><br/><br/><p>Large language models (LLMs) have improved so quickly <a href="https://spectrum.ieee.org/ai-math-benchmarks" target="_self">that the benchmarks themselves</a> have evolved, adding more complex problems in an effort to challenge the latest models. Yet LLMs haven’t improved across all domains, and one task remains far outside their grasp: They have no idea how to play video games. <br/><br/>While a few have managed to beat a few games (for example, <a href="https://techcrunch.com/2025/05/03/googles-gemini-has-beaten-pokemon-blue-with-a-little-help/" target="_blank">Gemini 2.5 Pro beat Pokemon Blue</a> in May of 2025), these exceptions prove the rule. The eventually victorious AI completed games far more slowly than a typical human player, made bizarre and often repetitive mistakes, and required custom software to guide their interactions with the game.</p><p><a href="http://julian.togelius.com/" rel="noopener noreferrer" target="_blank">Julian Togelius</a>, the director of New York University’s <a href="https://game.engineering.nyu.edu/" target="_blank">Game Innovation Lab</a> and co-founder of AI game-testing company Modl.ai, explored the implications of LLMs’ limitations in video games <a href="http://julian.togelius.com/Togelius2026What.pdf" target="_blank">in a recent paper</a>. He spoke with <em>IEEE Spectrum</em> about what this lack of video-game skills can tell us about the broader state of AI in 2026. <strong></strong></p><p><strong>LLMs have improved rapidly in coding, and your paper frames coding as a kind of well-behaved game. What do you mean by that?</strong> </p><p><strong>Julian Togelius: </strong>Coding is extremely well-behaved in the sense that you have tasks. These are like levels. You get a specification, you write code, and then you run it. <br/><br/>The reward is immediate and granular. The code has to compile, it has to run without crashing, and then it usually has to pass tests. Often, there’s also an explanation of how and why it failed. <br/><br/>There’s a theory from game designer <a href="https://www.raphkoster.com/" target="_blank">Raph Koster</a> that games are fun because we learn to play them as we play them. From that perspective, writing code is an extremely well-designed game. And in fact, writing code is something many people enjoy doing.<br/><br/><strong>Unlike coding, LLMs struggle with video games. This feels surprising given their <a data-linked-post="2671645555" href="https://spectrum.ieee.org/vibe-coding" target="_blank">success in coding</a>, as well as in games like chess and Go. What is it about video games that’s causing a problem?</strong> </p><p><strong>Togelius</strong>: It’s not just LLMs that are bad at this. We do not have general game AI.</p><p>There’s a widespread perception that because we can build AI that plays particular games well, we should be able to build one that plays any game. I’m not sure we’re going to get there. </p><p>People will mention that Google’s <a href="https://spectrum.ieee.org/deepmind-achieves-holy-grail" target="_self">AlphaZero</a> [which is not an LLM] can play both Go and chess. However, it had to be retrained and reengineered for each. And those are games that are similar in terms of input and output space. Most games are more different from each other. They have different mechanics and different input representations.</p><p>There’s also a data problem. Some of the games that AI can successfully play, like Minecraft and Pokémon, are among the most well-studied games in the world with literally millions of hours of guides. For a less well-known game, there’s far less. </p><h2>Video Game Benchmarks for LLM Performance</h2><p><strong>One factor that seems to help LLMs improve in coding is the proliferation of benchmarks. We have many benchmarks LLMs can try to solve, we can score the results, and then modify the LLM to improve performance. Developing a benchmark for playing a video game, though, is less clear-cut. Why is that?</strong></p><p><strong>Togelius: </strong>I’ve built many game-based AI benchmarks over the years. One, <a href="https://cdn.aaai.org/ojs/9869/9869-13-13397-1-2-20201228.pdf" rel="noopener noreferrer" target="_blank">the General Video Game AI competition</a>, ran for seven years. We tested an agent on our publicly available games, and every time we ran the competition, we invented 10 new games to test on. <br/><br/>One reason we stopped was that we stopped seeing progress. Agents got better at some games but worse at others. This was before LLMs.<br/><br/>Lately, we’ve been updating this framework for LLMs. They fail. They absolutely suck. All of them. They don’t even do as well as a simple search algorithm. <br/><br/>Why? They were never trained on these games, and they’re separately very bad at spatial reasoning. Which shouldn’t be surprising, because that’s also not in the training data.</p><p><strong>This brings us to what seems like a contradiction. LLMs are bad at playing games. Yet at the same time, they’re improving rapidly at coding, a skill set that can be used to create a game. How do these facts fit together?</strong> </p><p><strong>Togelius: </strong>It’s super weird. You can go into Cursor or Claude, write one prompt, and get a playable game. The game will be very typical, because an LLM’s code-writing abilities are better the more typical something is. So, if you ask it to give you something like <a data-linked-post="2656808350" href="https://spectrum.ieee.org/commodore-64" target="_blank">Asteroids</a>, it will work. That’s impressive.<br/><br/>However, it’s not going to give you a good or novel game. That does seem weird. The reason is that the LLM can’t play it. Game development is an iterative process. You write, you test, you adjust the game feel. An LLM can’t do that. </p><p>And to an extent, I don’t think it’s different when designing other software. Yes, you can ask an LLM to create a GUI with a bunch of buttons. But the LLM doesn’t know much about how to use it. </p><p><strong>Companies like Nvidia and Google have talked about using simulations, including gamelike environments, to improve AI performance. If AI can’t master games in general, how optimistic should we be about that approach?</strong></p><p><strong>Togelius: </strong>Games are both easier and harder than the real world. They’re easier because there are fewer levels of abstraction. They’re harder because games are much more diverse. The real world has the same physics everywhere. </p><p>One example is Waymo, which uses world models in its training loop. That makes sense because driving is much the same everywhere. It’s way less diverse than games. </p><p>That’s confusing for people. People see an LLM write an academic essay on quantum physics and wonder, “How can it not play both Halo and Space Invaders?” However, those games are more different from each other, in a sense, than two academic essays. </p>]]></description><pubDate>Sun, 29 Mar 2026 13:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/ai-video-games-llms-togelius</guid><category>Llms</category><category>Artificial-intelligence</category><category>Video-games</category><dc:creator>Matthew S. Smith</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-middle-aged-man-smiling-while-holding-a-video-game-controller-behind-him-is-a-bookshelf-filled-with-countless-cooperative-boa.jpg?id=65413486&amp;width=980"></media:content></item><item><title>NYU’s Quantum Institute Bridges Science and Application</title><link>https://spectrum.ieee.org/nyu-quantum-institute</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/person-in-white-suit-working-with-semiconductor-equipment-in-a-lab.jpg?id=65322091&width=1200&height=800&coordinates=350%2C0%2C350%2C0"/><br/><br/><p><em>This sponsored article is brought to you by <a href="https://engineering.nyu.edu/" rel="noopener noreferrer" target="_blank">NYU Tandon School of Engineering</a>.</em></p><p>Within a 6 mile radius of New York University’s (NYU) campus, there are more than 500 tech industry giants, banks, and hospitals. This isn’t just a fact about real estate, it’s the foundation for advancing quantum discovery and application.</p><p>While the world races to harness quantum technology, NYU is betting that the ultimate advantage lies not solely in a lab, but in the dense, demanding, and hyper-connected urban ecosystem that surrounds it. With the launch of its <a href="https://www.nyu.edu/about/news-publications/news/2025/october/nyu-launches-quantum-institute-.html" rel="noopener noreferrer" target="_blank"><span>NYU Quantum Institute</span></a> (NYUQI), NYU is positioning itself as <a href="https://www.nyu.edu/about/news-publications/news/2025/october/top-quantum-scientists-convene-at-nyu.html" target="_blank">the central node</a> in this network; a “full stack” powerhouse built on the conviction that it has found the right place, and the right time, to turn quantum science into tangible reality.</p><p>Proximity advantage is essential because quantum science demands it. Globally, the quest for practical quantum solutions — whether for computing, sensing, or secure communications — has been stalled, in part, by fragmentation. Physicists and chemical engineers invent new materials, computer scientists develop new algorithms, and electrical engineers build new devices, but all three often work in isolated academic silos.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Three men pose at the 4th Annual NYC Quantum Summit 2025; attendees converse in the background." class="rm-shortcode" data-rm-shortcode-id="1dd6dfe45b73630bb9040545fcdfae7d" data-rm-shortcode-name="rebelmouse-image" id="33e2d" loading="lazy" src="https://spectrum.ieee.org/media-library/three-men-pose-at-the-4th-annual-nyc-quantum-summit-2025-attendees-converse-in-the-background.jpg?id=65322345&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Gregory Gabadadze, NYU’s dean for science, NYU physicist and Quantum Institute Director Javad Shabani, and Juan de Pablo, Anne and Joel Ehrenkranz Executive Vice President for Global Science and Technology and executive dean of the Tandon School of Engineering.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Veselin Cuparić/NYU</small></p><p><span>NYUQI’s premise is that breakthroughs happen “at the interfaces between different domains,” according to </span><a href="https://engineering.nyu.edu/faculty/juan-de-pablo" target="_blank"><span>Juan de Pablo</span></a><span>, Executive Vice President for Global Science and Technology at NYU and Executive Dean of the NYU Tandon School of Engineering. The Institute is built to actively force those necessary collisions — to integrate the physicists, engineers, materials scientists, computer scientists, biologists, and chemists vital to quantum research into one holistic operation. This institutional design ensures that the hardware built by one team can be immediately tested by software developed by another, accelerating progress in a way that isolated departments never could.</span></p><p class="pull-quote"><span>NYUQI’s premise is that breakthroughs happen at the interfaces between different domains. <strong>—Juan de Pablo, NYU Tandon School of Engineering</strong></span></p><p>NYUQI’s integrated vision is backed by a massive physical commitment to the city. The NYUQI is not just a theoretical concept; its collaborators will be housed in a renovated, <a href="https://www.nyu.edu/about/news-publications/news/2025/may/nyu-entering-long-term-lease-at-770-broadway.html" target="_blank"><span>million-square-foot facility</span></a> in the heart of Manhattan’s West Village, backed by a state-of-the-art <a href="https://engineering.nyu.edu/research/nanofab" target="_blank">Nanofabrication Cleanroom</a> in Brooklyn serving as a high-tech foundry. This is where the theoretical meets physical devices, allowing the Institute to test and refine the process from materials science to deployment.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt='NYU building exterior with "Science + Tech" signage, flags, and a passing yellow taxi.' class="rm-shortcode" data-rm-shortcode-id="605cc71d844927d3fb0a05fb086fedcf" data-rm-shortcode-name="rebelmouse-image" id="bceaa" loading="lazy" src="https://spectrum.ieee.org/media-library/nyu-building-exterior-with-science-tech-signage-flags-and-a-passing-yellow-taxi.jpg?id=65322352&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">NYUQI will be housed in a renovated, million-square-foot facility in the heart of Manhattan’s West Village.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Tracey Friedman/NYU</small></p><p><span>Leading this effort is NYUQI Director </span><a href="https://as.nyu.edu/faculty/javad-shabani.html" target="_blank"><span>Javad Shabani</span></a><span>, who, along with the other members, is turning the Institute into a hub for collaboration with private and public sector partners with quantum challenges that need solving. As de Pablo explains, “Anybody who wants to work on quantum with NYU, you come in through that door, and we’ll send you to the right place.” For New York’s vast ecosystem of tech giants and financial institutions, the NYUQI offers a resource they can’t build on their own: a cohesive team of experts in quantum phenomena, quantum information theory, communication, computing, materials, and optics, and a structured path to applying theoretical discoveries to advanced quantum technologies.</span></p><h2>Solving the Challenge of Quantum Research</h2><p><span>The NYUQI’s integrated structure is less about organizational management, and more about scientific requirement. </span><span>The challenge of quantum is that the hardware, the software, and the programming are inherently interconnected — each must be designed to work with the other. To solve this, the Institute focuses on three applications of quantum science: Quantum Computing, Quantum Sensing, and Quantum Communications.</span></p><p>For Shabani, this means creating an integrated environment that bridges discovery with experimentation, starting with the physical components all the way to quantum algorithm centers. That will include a fabrication facility in the new building in Manhattan, as well as the <a href="https://engineering.nyu.edu/news/chips-and-science-act-spurs-nanofab-cleanroom-ribbon-cutting-nyu-tandon-school-engineering" target="_blank"><span>NYU Nanofab</span></a> in Brooklyn directed by Davood Shahjerdi. New York Senators Charles Schumer and Kirsten Gillibrand recently secured <a href="https://www.nyu.edu/about/news-publications/news/2026/february/nyu-receives--1-million-in-funding-from-senators-schumer-and-gil.html" target="_blank">$1 million in congressionally-directed spending</a> to bring Thermal Laser Epitaxy (TLE) technology — which allows for atomic-level purity, minimal defects, and streamlined application of a diverse range of quantum materials — to NYU, marking the first time the equipment will be used in the U.S.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Two people hold semiconductor wafers during a presentation with audience taking photos." class="rm-shortcode" data-rm-shortcode-id="1a0dbca6c6bb8fb7dbf4d399689b2922" data-rm-shortcode-name="rebelmouse-image" id="d434c" loading="lazy" src="https://spectrum.ieee.org/media-library/two-people-hold-semiconductor-wafers-during-a-presentation-with-audience-taking-photos.jpg?id=65322354&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">NYU Nanofab manager Smiti Bhattacharya and Nanofab Director Davood Shahjerdi at the nanofab ribbon-cutting in 2023. The nanofab is the first academic cleanroom in Brooklyn, and serves as a prototyping facility for the NORDTECH Microelectronics Commons consortium.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">NYU WIRELESS</small></p><p>Tight control over fabrication, and can allow researchers to pivot quickly when a breakthrough in one area — say, finding a cheaper, more reliable material like silicon carbide — can be explored for use across all three applications, and offers unique access to academics and the private sector alike to sophisticated pieces of specialty equipment whose maintenance knowledge and costs make them all-but-impossible to maintain outside of the right staffing and environment.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="3D model of a laboratory layout, highlighting the Yellow Room in bright yellow." class="rm-shortcode" data-rm-shortcode-id="e7c1128703d96de919ed2ce440a97416" data-rm-shortcode-name="rebelmouse-image" id="62d58" loading="lazy" src="https://spectrum.ieee.org/media-library/3d-model-of-a-laboratory-layout-highlighting-the-yellow-room-in-bright-yellow.png?id=65322596&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">The NYU Nanofab is Brooklyn’s first academic cleanroom, with a strategic focus on superconducting quantum technologies, advanced semiconductor electronics, and devices built from quantum heterostructures and other next-generation materials.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">NYU Nanofab</small></p><p><span>That speed and adaptability is the NYUQI’s competitive edge. It turns fragmented challenges into holistic solutions, positioning the Institute to solve real-world problems for its New York neighbors—from highly secure data transmission to next-generation drug discovery.</span></p><h2>Testing Quantum Communication in NYC</h2><p>The integrated approach also makes the NYUQI a testbed for the most critical near-term applications. Take Quantum Communications, which is essential for creating an “unhackable” quantum internet. In an industry first, NYU worked with the quantum start-up Qunnect to <a href="https://www.nyu.edu/about/news-publications/news/2023/september/nyu-takes-quantum-step-in-establishing-cutting-edge-tech-hub-in-.html" target="_blank"><span>send quantum information through standard telecom fiber</span></a> in New York City between Manhattan and Brooklyn through a 10-mile quantum networking link. Instead of simulating communication challenges in a lab, the NYUQI team is already leveraging NYU’s city-wide campus by utilizing existing infrastructure to test secure quantum transmission between Manhattan and Brooklyn. </p><p class="pull-quote">The NYUQI team is already leveraging NYU’s city-wide campus by utilizing existing infrastructure to test secure quantum transmission between Manhattan and Brooklyn.</p><p>This isn’t just theory; it is building a functioning prototype in the most demanding, dense urban environment  in the world. Real-time, real-world deployment is a critical component missing in other isolated institutions. When the NYUQI achieves results, the technology will be that much more readily available to the massive financial, tech, and communications organizations operating right outside their door.</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="Scientist in protective gear working in a laboratory with samples." class="rm-shortcode" data-rm-shortcode-id="d644b791788af64769a853d0516834e6" data-rm-shortcode-name="rebelmouse-image" id="dc2fb" loading="lazy" src="https://spectrum.ieee.org/media-library/scientist-in-protective-gear-working-in-a-laboratory-with-samples.jpg?id=65322378&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">NYUQI includes a state-of-the-art Nanofabrication Cleanroom in Brooklyn serving as a high-tech foundry.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">NYU Tandon</small></p><p><span>While the Institute has built the physical infrastructure and designed the necessary scientific architecture, its enduring contribution will be the specialized workforce it creates for the new quantum economy. This addresses the market’s greatest deficit: a lack of individuals trained not just in physics, but in the integrated, full-stack approach that quantum demands.</span></p><p>By creating a pipeline of 100 to 200 graduate and doctoral students who are encouraged to collaborate across Computing, Sensing, and Communications, the NYUQI is narrowing the skills gap. These will be future leaders who can speak the language of the physicist, the materials scientist, and the engineer simultaneously. This commitment to interdisciplinary talent is also fueled by the launch of the new Master of Science in Quantum Science & Technology program at NYU Tandon, positioning the university among a select group worldwide offering such a specialized degree.</p><p>Interdisciplinary education creates the shared language and understanding poised to make graduates coming from collaborations in the NYUQI extremely valuable in the current landscape. Quantum challenges are not just technical; they are managerial and philosophical as well. An engineer working with the NYUQI will understand the requirements of the nanofabrication cleanroom and the foundations of superconducting qubits for quantum computing, just as a physicist will understand the application needs of an industry partner like a large financial institution. In a field where the entire team must be able to communicate seamlessly, these are professionals truly equipped to rapidly translate discovery into deployable technology. Creating a talent pipeline at scale will provide a missing link that converts New York’s vast commercial energy into genuine quantum advantage.</p><h2>NYUQI: Building Talent, Technology, and Structure</h2><p><span>The vision for the NYUQI </span><span>is an act of strategic geography that plays directly into the sheer volume of opportunity and demand right outside their new facility. </span><span>By building the talent, the technology, and the structure necessary to capitalize on this dense environment, NYU is not just participating in the quantum race, it is actively steering it.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Conference room with attendees seated at round tables, facing a presenter on stage." class="rm-shortcode" data-rm-shortcode-id="f5e2ae16e0c5ebc4f0828d52ed639115" data-rm-shortcode-name="rebelmouse-image" id="02b7e" loading="lazy" src="https://spectrum.ieee.org/media-library/conference-room-with-attendees-seated-at-round-tables-facing-a-presenter-on-stage.jpg?id=65322370&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Attendees of NYU’s 2025 Quantum Summit.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Tracey Friedman/NYU</small></p><p>The initial hypothesis for the NYUQI was simple: the ultimate advantage lies in pursuing the science in the right place at the right time. Now, the institute will ensure that the next wave of scientific discovery, capable of solving previously intractable problems in finance, medicine, and security, will be conceived, built, and tested in the heart of New York City.</p>]]></description><pubDate>Fri, 27 Mar 2026 10:02:05 +0000</pubDate><guid>https://spectrum.ieee.org/nyu-quantum-institute</guid><category>Nyu-tandon</category><category>Quantum-computing</category><category>Quantum-internet</category><category>Semiconductors</category><category>Quantum-communications</category><dc:creator>Wiley</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/person-in-white-suit-working-with-semiconductor-equipment-in-a-lab.jpg?id=65322091&amp;width=980"></media:content></item><item><title>Training Driving AI at 50,000× Real Time</title><link>https://spectrum.ieee.org/gm-scalable-driving-ai</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/sleek-suv-driving-on-a-highway-surrounded-by-trees-under-a-clear-blue-sky.png?id=65321052&width=1200&height=800&coordinates=156%2C0%2C156%2C0"/><br/><br/><p><em>This is a sponsored article brought to you by General Motors. Visit their new </em><em><a href="https://engineering.gm.com/home.html" rel="noopener noreferrer" target="_blank">Engineering Blog</a></em><em> for more insights.</em></p><p>Autonomous driving is one of the most demanding problems in physical AI. An automated system must interpret a chaotic, ever-changing world in real time—navigating uncertainty, predicting human behavior, and operating safely across an immense range of environments and edge cases.</p><p>At General Motors, we approach this problem from a simple premise: while most moments on the road are predictable, the rare, ambiguous, and unexpected events — the long tail — are what ultimately defines whether an autonomous system is safe, reliable, and ready for deployment at scale. (Note: While here we discuss research and emerging technologies to solve the long tail required for full general autonomy, we also discuss our current approach or solving 99% of everyday autonomous driving in a deep dive on Compound AI.)</p><p>As GM advances toward <a href="https://news.gm.com/home.detail.html/Pages/news/us/en/2025/oct/1022-AI-GM-launch-eyes-off-driving-conversational-AI.html" target="_blank">eyes-off</a> highway driving, and ultimately toward fully autonomous vehicles, solving the long tail becomes the central engineering challenge. It requires developing systems that can be counted on to behave sensibly in the most unexpected conditions.</p><p>GM is <a href="https://neurips.cc/virtual/2025/loc/san-diego/128661" rel="noopener noreferrer" target="_blank">building scalable driving AI</a> to meet that challenge — combining large-scale simulation, reinforcement learning, and foundation-model-based reasoning to train autonomous systems at a scale and speed that would be impossible in the real world alone.</p><h2>Stress-testing for the long tail</h2><p>Long-tail scenarios of autonomous driving come in a few varieties.</p><p>Some are notable for their rareness. There’s a mattress on the road. A fire hydrant bursts. A massive power outage in San Francisco that disabled traffic lights required <a href="https://waymo.com/blog/2025/12/autonomously-navigating-the-real-world" rel="noopener noreferrer" target="_blank">driverless vehicles to navigate</a> never-before experienced challenges. These rare system-level interactions, especially in dense urban environments, show how unexpected edge cases can cascade at scale.</p><p>But long-tail challenges don’t just come in the form of once-in-a-lifetime rarities. They also manifest as everyday scenarios that require characteristically human courtesy or common sense. How do you queue up for a spot without blocking traffic in a crowded parking lot? Or navigate a construction zone, guided by gesturing workers and ad-hoc signs? These are simple challenges for a human driver but require inventive engineering to handle flawlessly with a machine.</p><h3>Autonomous driving scenario demand curve</h3><br/><img alt="Graph showing scenario complexity: Predictable, everyday, and rare long-tail events." class="rm-shortcode" data-rm-shortcode-id="363e07d5da15d3b590d5cf1f9c13ba02" data-rm-shortcode-name="rebelmouse-image" id="a0363" loading="lazy" src="https://spectrum.ieee.org/media-library/graph-showing-scenario-complexity-predictable-everyday-and-rare-long-tail-events.png?id=65321037&width=980"/><h3></h3><br/><h2>Deploying vision language models</h2><p>One tool GM is developing to tackle these nuanced scenarios is the use of Vision Language Action (VLA) models. Starting with a standard Vision Language Model, which leverages internet-scale knowledge to make sense of images, GM engineers use specialized decoding heads to fine-tune for distinct driving-related tasks. The resulting VLA can make sense of vehicle trajectories and detect 3D objects on top of its general image-recognition capabilities.</p><p>These tuned models enable a vehicle to recognize that a police officer’s hand gesture overrides a red traffic light or to identify what a “loading zone” at a busy airport terminal might look like.</p><p>These models can also generate reasoning traces that help engineers and safety operators understand why a maneuver occurred — an important tool for debugging, validation, and trust.</p><h2>Testing hazardous scenarios in high-fidelity simulations</h2><p>The trouble is: driving requires split-second reaction times so any excess latency poses an especially critical problem. To solve this, GM is developing a “Dual Frequency VLA.” This large-scale model runs at a lower frequency to make high-level semantic decisions (“Is that object in the road a branch or a cinder block?”), while a smaller, highly efficient model handles the immediate, high-frequency spatial control (steering and braking).</p><p>This hybrid approach allows the vehicle to benefit from deep semantic reasoning without sacrificing the split-second reaction times required for safe driving.</p><p>But dealing with an edge case safely requires that the model not only understand what it is looking at but also understand how to sensibly <em>drive through</em> the challenge it’s identified. For that, there is no substitute for experience.</p><p>Which is why, each day, <a href="https://news.gm.com/home.detail.html/Pages/topic/us/en/2025/oct/1009-GMs-path-full-autonomy-Building-trust-step-by-step.html%29" rel="noopener noreferrer" target="_blank">we run millions of high-fidelity closed loop simulations</a>, equivalent to tens of thousands of human driving days, compressed into hours of simulation. We can replay actual events, modify real-world data to create new virtual scenarios, or design new ones entirely from scratch. This allows us to regularly test the system against hazardous scenarios that would be nearly impossible to encounter safely in the real world.</p><h2>Synthetic data for the hardest cases</h2><p>Where do these simulated scenarios come from? GM engineers employ a whole host of AI technologies to produce novel training data that can model extreme situations while remaining grounded in reality.</p><p>GM’s <a href="https://bmvc2025.bmva.org/proceedings/154/" rel="noopener noreferrer" target="_blank">“Seed-to-Seed Translation” research</a>, for instance, leverages diffusion models to transform existing real-world data, allowing a researcher to turn a clear-day recording into a rainy or foggy night while perfectly preserving the scene’s geometry. The result? A “domain change”—clear becomes rainy, but everything else remains the same.</p><p>In addition, our GM World diffusion-based simulator allows us to synthesize entirely new traffic scenarios using natural language and spatial bounding boxes. We can summon entirely new scenarios with different weather patterns. We can also take an existing road scene and add challenging new elements, such as a vehicle cutting into our path.</p><h3></h3><br/><img alt='Comparison of a 3D model and street view with a vehicle removed, labeled "Original" and "Edited".' class="rm-shortcode" data-rm-shortcode-id="0d17abe7a8791b531b5951439023ffa9" data-rm-shortcode-name="rebelmouse-image" id="f3e89" loading="lazy" src="https://spectrum.ieee.org/media-library/comparison-of-a-3d-model-and-street-view-with-a-vehicle-removed-labeled-original-and-edited.gif?id=65321060&width=980"/><h3></h3><br/><img alt="Street with several cars parked, partially flooded after heavy rain; blue geometric markings overlay." class="rm-shortcode" data-rm-shortcode-id="79227a75d41c9ee4e257dd3cd21a80e7" data-rm-shortcode-name="rebelmouse-image" id="c1ebf" loading="lazy" src="https://spectrum.ieee.org/media-library/street-with-several-cars-parked-partially-flooded-after-heavy-rain-blue-geometric-markings-overlay.gif?id=65321061&width=980"/><h3></h3><br/><img alt="Winter street with cars; blue 3D wireframe shapes overlay." class="rm-shortcode" data-rm-shortcode-id="94864ff674e3b311d384ec0114587d8d" data-rm-shortcode-name="rebelmouse-image" id="55554" loading="lazy" src="https://spectrum.ieee.org/media-library/winter-street-with-cars-blue-3d-wireframe-shapes-overlay.gif?id=65321063&width=980"/><h2></h2><p>High-fidelity simulation isn’t always the best tool for every learning task. Photorealistic rendering is essential for training perception systems to recognize objects in varied conditions. But when the goal is teaching decision-making and tactical planning—when to merge, or how to navigate an intersection—the computationally expensive details matter less than spatial relationships and traffic dynamics. AI systems may need billions or even trillions of lightweight examples to support reinforcement learning, where models learn the rules of sensible driving through rapid trial and error rather than relying on imitation alone.</p><p>To this end, General Motors has developed a proprietary, multi-agent reinforcement learning simulator, GM Gym, to serve as a closed-loop simulation environment that can both simulate high-fidelity sensor data, and model thousands of drivers per second in an abstract environment known as “Boxworld.”</p><p>By focusing on essentials like spatial positioning, velocity and rules of the road while stripping away details like puddles and potholes, Boxworld creates a high-speed training environment for reinforcement learning models at incredible speeds, operating 50,000 times faster than real-time and simulating 1,000 km of driving per second of GPU time. It’s a method that allows us to not just imitate humans, but to develop driving models that have verifiable objective outcomes, like safety and progress.</p><h2>From abstract policy to real-world driving</h2><p>Of course, the route from your home to your office does not run through Boxworld. It passes through a world of asphalt, shadows, and weather. So, to bring that conceptual expertise into the real world, GM is one of the first to employ a technique called “On Policy Distillation,” where engineers run their simulator in both modes simultaneously: the abstract, high-speed Boxworld and the high-fidelity sensor mode.</p><p>Here, the reinforcement learning model—which has practiced countless abstract miles to develop a perfect “policy,” or driving strategy—acts as a teacher. It guides its “student,” the model that will eventually live in the car. This transfer of wisdom is incredibly efficient; just 30 minutes of distillation can capture the equivalent of 12 hours of raw reinforcement learning, allowing the real-world model to rapidly inherit the safety instincts its cousin painstakingly honed in simulation.</p><h2>Designing failures before they happen</h2><p>Simulation isn’t just about training the model to drive well, though; it’s also about trying to make it fail. To rigorously stress-test the system, GM utilizes <a href="https://arxiv.org/abs/2309.05810" target="_blank">a differentiable pipeline called SHIFT3D</a>. Instead of just recreating the world, SHIFT3D actively modifies it to create “adversarial” objects designed to trick the perception system. The pipeline takes a standard object, like a sedan, and subtly morphs its shape and pose until it becomes a “challenging”, fun-house version that is harder for the AI to detect. Optimizing these failure modes is what allows engineers to preemptively discover safety risks before they ever appear on the road. Iteratively retraining the model on these generated “hard” objects has been shown to reduce near-miss collisions by over 30%, closing the safety gap on edge cases that might otherwise be missed.</p><p>Even with advanced simulation and adversarial testing, a truly robust system must know its own limits. To enable safety in the face of the unknown, GM researchers add a specialized “Epistemic uncertainty head” to their models. This architectural addition allows the AI to distinguish between standard noise and genuine confusion. When the model encounters a scenario it doesn’t understand—a true “long tail” event—it signals high epistemic uncertainty. This acts as a principled proxy for data mining, automatically flagging the most confusing and high-value examples for engineers to analyze and add to the training set.</p><p>This rigorous, multi-faceted approach—from “Boxworld” strategy to adversarial stress-testing—is General Motors’ proposed framework for solving the final 1% of autonomy. And while it serves as the foundation for future development, it also surfaces new research challenges that engineers must address.</p><p>How do we balance the essentially unlimited data from Reinforcement Learning with the finite but richer data we get from real-world driving? How close can we get to full, human-like driving by writing down a reward function? Can we go beyond domain change to generate completely new scenarios with novel objects?</p><h2>Solving the long tail at scale</h2><p>Working toward solving the long tail of autonomy is not about a single model or technique. It requires an ecosystem — one that combines high-fidelity simulation with abstract learning environments, reinforcement learning with imitation, and semantic reasoning with split-second control.</p><p>This approach does more than improve performance on average cases. It is designed to surface the rare, ambiguous, and difficult scenarios that determine whether autonomy is truly ready to operate without human supervision.</p><p>There are still open research questions. How human-like can a driving policy become when optimized through reward functions? How do we best combine unlimited simulated experience with the richer priors embedded in real human driving? And how far can generative world models take us in creating meaningful, safety-critical edge cases?</p><p>Answering these questions is central to the future of autonomous driving. At GM, we are building the tools, infrastructure, and research culture needed to address them — not at small scale, but at the scale required for real vehicles, real customers, and real roads.</p>]]></description><pubDate>Wed, 25 Mar 2026 19:00:05 +0000</pubDate><guid>https://spectrum.ieee.org/gm-scalable-driving-ai</guid><category>Autonomous-vehicles</category><category>Self-driving-cars</category><category>Gm</category><dc:creator>Ben Snyder</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/sleek-suv-driving-on-a-highway-surrounded-by-trees-under-a-clear-blue-sky.png?id=65321052&amp;width=980"></media:content></item><item><title>What Happens When You Host an AI Café</title><link>https://spectrum.ieee.org/ai-community-engagement</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/hands-hold-a-coffee-cup-with-the-letters-ai-in-white-decorative-foam.jpg?id=65351357&width=1200&height=800&coordinates=0%2C1%2C0%2C1"/><br/><br/><p>“Can I get an interview?” “Can I get a job when I graduate?” Those questions came from students during a candid discussion about artificial intelligence, capturing the anxiety many young people feel today. As companies adopt AI-driven interview screeners, restructure their workforces, and redirect billions of dollars toward <a href="https://spectrum.ieee.org/ai-data-centers-engineers-jobs" target="_blank">AI infrastructure</a>, students are increasingly unsure of what the future of work will look like.</p><p>We had gathered people together at a coffee shop in Auburn, Alabama, for what we called an AI Café. The event was designed to confront concerns about AI directly, demystifying the technology while pushing back against the growing narrative of technological doom. </p><p>AI is reshaping society at breathtaking speed. Yet the trajectory of this transformation is being charted primarily by for-profit tech companies, whose priorities revolve around market dominance rather than public welfare. Many people feel that AI is something being done <em><em>to</em></em> them rather than developed <em><em>with</em></em> them.</p><p>As computer science and liberal arts faculty at <a href="https://www.auburn.edu/" target="_blank">Auburn University</a>, we believe there is another path forward: one where scholars engage their communities in genuine dialogue about AI. Not to lecture about technical capabilities, but to listen, learn, and co-create a vision for AI that serves the public interest.</p><h2>The AI Café Model</h2><p>Last November, we ran<strong> </strong>two public <a href="https://cla.auburn.edu/news/articles/auburn-faculty-lead-community-conversations-about-ai/" target="_blank">AI Cafés</a> in Auburn. These were informal, 90-minute conversations between faculty, students, and community members about their experiences with AI.<strong> </strong>In these conversational forums, participants sat in clusters, questions flowed in multiple directions, and lived experience carried as much weight as technical expertise.</p><p>We avoided jargon and resisted attempts to “correct” misconceptions, welcoming whatever emotions emerged. One ground rule proved crucial: keeping discussions in the present, asking participants where they encounter AI today. Without that focus, conversations could easily drift to <a href="https://spectrum.ieee.org/artificial-general-intelligence" target="_blank">sci-fi speculation</a>. Historical analogies—to the printing press, electricity, and smartphones—helped people contextualize their reactions. And we found that without shared definitions of AI, people talked past each other; we learned to ask participants to name specific tools they were concerned about.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A pair of photos show people in chairs in a cafe raising their hands, and 3 people smiling in front of the audience." class="rm-shortcode" data-rm-shortcode-id="f35dab7bb7c94eb3c1ec083a27997de2" data-rm-shortcode-name="rebelmouse-image" id="2956f" loading="lazy" src="https://spectrum.ieee.org/media-library/a-pair-of-photos-show-people-in-chairs-in-a-cafe-raising-their-hands-and-3-people-smiling-in-front-of-the-audience.jpg?id=65352141&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Organizers Xaq Frohlich, Cheryl Seals, and Joan Harrell (right) held their first AI Café in a welcoming coffee shop and bookstore. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit..."><a href="https://www.wellredau.com/" target="_blank">Well Red</a></small></p><p>Most important, we approached these events not as experts enlightening the masses, but as community members navigating complex change together.</p><h2>What We Learned by Listening</h2><p>Participants arrived with significant frustration. They felt that commercial interests were driving AI development “without consideration of public needs,” as one attendee put it. This echoed deeper anxieties about technology, from <a href="https://spectrum.ieee.org/tag/social-media" target="_blank">social media</a> algorithms that amplify division to devices that profit from “engagement” and replace meaningful face-to-face connection. People aren’t simply “afraid of AI.” They’re weary of a pattern where powerful technologies reshape their lives while they have little say.</p><p>Yet when given space to voice concerns without dismissal, something shifted. Participants didn’t want to stop AI development; they wanted to have a voice in it. When we asked “What would a human-centered AI future look like?” the conversation became constructive. People articulated priorities: fairness over efficiency, creativity over automation, dignity over convenience, community over individualism.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Three people standing together in front of a yellow curtain at an indoor event." class="rm-shortcode" data-rm-shortcode-id="26cf47b8431459d9c9ed0bf5069d1f90" data-rm-shortcode-name="rebelmouse-image" id="db5c6" loading="lazy" src="https://spectrum.ieee.org/media-library/three-people-standing-together-in-front-of-a-yellow-curtain-at-an-indoor-event.jpg?id=65357899&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">The three organizers, all professors at Alabama’s Auburn University, say that including people from the liberal arts fields brought new perspectives to the discussions about AI. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit..."><a href="https://www.wellredau.com/" target="_blank">Well Red</a></small></p><p>For us as organizers, the experience was transformative. Hearing how AI affected people’s work, their children’s education, and their trust in information prompted us to consider dimensions we hadn’t fully grasped. Perhaps most striking was the gratitude participants expressed for being heard. It wasn’t about filling knowledge deficits; it was about mutual learning. The trust generated created a spillover effect, renewing faith that AI could serve the public interest if shaped through inclusive processes.</p><h2>How to Start Your Own AI Café</h2><p>The “deficit model” of science communication—where experts transmit knowledge to an uninformed public—has been discredited. Public resistance to emerging technologies reflects legitimate concerns about values, risks, and who controls decision-making. Our events point toward a better model.</p><p>We urge engineering and liberal arts departments, professional societies, and community organizations worldwide to organize dialogues similar to our AI Cafés.</p><p>We found that a few simple design choices made these conversations far more productive.<strong> </strong>Informal and welcoming spaces such as coffee shops, libraries, and community centers helped participants feel comfortable (and serving food and drinks helped too!). Starting with small-group discussions, where<strong> </strong>people talked with neighbors, produced more honest thinking and greater participation. Partnering with colleagues in the liberal arts brought additional perspectives on technology’s social dimensions. And by making a commitment to an ongoing series of events, we built trust.</p><p>Facilitation also matters. Rather than leading with technical expertise, we began with values: We asked what kind of world participants wanted, and how AI might help or hinder that vision. We used analogies to earlier technologies to help people situate their reactions and grounded discussions in present realities, asking participants where they have encountered AI in their daily lives. We welcomed emotions constructively, transforming worry into problem solving by<strong> </strong>asking questions like: “What would you do about that?”</p><h2>Why Engineers Should Engage the Public</h2><p>Professional <a href="https://techethics.ieee.org/" target="_blank">ethics codes</a> remain abstract unless grounded in dialogue with affected communities. Conversations about what “responsible AI” means will look different in São Paulo than in Seoul, in Vienna than in Nairobi. What makes the AI Café model portable is its general principles: informal settings, values-first questions, present-tense focus, genuine listening.</p><p>Without such engagement, ethical accountability quietly shifts to technical experts rather than remaining a shared public concern. If we let commercial interests define AI’s trajectory with minimal public input, it will only deepen divides and <a href="https://spectrum.ieee.org/joy-buolamwini/joy-buolamwini" target="_blank">entrench inequities</a>.</p><p>AI will continue advancing whether or not we have public trust. But AI shaped through dialogue with communities will look fundamentally different from AI developed solely to pursue what’s technically possible or commercially profitable.</p><p>The tools for this work aren’t technical; they’re social, requiring humility, patience, and genuine curiosity. The question isn’t whether AI will transform society. It’s whether that transformation will be done <em><em>to</em></em> people or <em><em>with</em></em> them. We believe scholars must choose the latter, and that starts with showing up in coffee shops and community centers to have conversations where we do less talking and more listening.</p><p>The future of AI depends on it.</p><em><em><br/></em></em>]]></description><pubDate>Wed, 25 Mar 2026 14:00:05 +0000</pubDate><guid>https://spectrum.ieee.org/ai-community-engagement</guid><category>Ethics</category><category>Community-values</category><category>Responsible-ai</category><category>Algorithmic-bias</category><dc:creator>Xaq Frohlich</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/hands-hold-a-coffee-cup-with-the-letters-ai-in-white-decorative-foam.jpg?id=65351357&amp;width=980"></media:content></item><item><title>These AI Workstations Look Like PCs but Pack a Stronger Punch</title><link>https://spectrum.ieee.org/ai-workstation-looks-like-pcs</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-sleek-illuminated-box-resembling-a-personal-computer-workstation.jpg?id=65329122&width=1200&height=800&coordinates=0%2C208%2C0%2C209"/><br/><br/><p>The rise of generative AI has spurred demand for AI workstations that can run or train models on local hardware. Yet modern PCs <a href="https://spectrum.ieee.org/ai-models-locally" target="_self">have proven inadequate for this task</a>. A typical laptop has only enough memory to load a large language model (LLM) with 8 billion to 13 billion parameters—much smaller, and much less intelligent, than frontier models that are presumed to have over a trillion parameters. Even the most capable workstation PCs struggle to serve LLMs with more than 70 billion parameters.</p><p><a href="https://tenstorrent.com/waitlist/tt-quietbox" rel="noopener noreferrer" target="_blank">Tenstorrent’s QuietBox 2</a> is an attempt to fill that gap. Though it looks like a PC workstation, the QuietBox 2 contains four of the company’s custom Blackhole AI accelerators, 128 gigabytes of GDDR6 memory—specialized memory used in GPUs—and 256 GB of DDR5 system memory (for a total of 384 GB). This configuration provides enough memory to load OpenAI’s GPT-OSS-120B and can run <a href="https://spectrum.ieee.org/meta-llama-3" target="_blank">midsize models</a> like Meta’s Llama 3.1 70B at speeds of nearly 500 tokens per second. For reference, that’s several times quicker than an average response from OpenAI’s GPT-5.2 or Anthropic’s Claude 4.6. The QuietBox 2 carries an expected retail price of US $9,999 and is slated to launch in the second quarter of 2026. </p><p>“The 128 gigabytes of GDDR that we have with our AI accelerators really defines how big of a model you can run at a reasonable speed,” says <a href="https://www.linkedin.com/in/milosstrajkovic/?originalSubdomain=ca" rel="noopener noreferrer" target="_blank">Milos Trajkovic</a>, cofounder and systems engineer at Tenstorrent. “Our 128 gigabytes of GDDR6 RAM would require four Nvidia RTX 5090 graphics cards. That couldn’t fit in today’s 1,600-watt form factor, and the cost for four RTX 5090 GPUs is huge.”</p><h2>An AI workstation built at the home office</h2><p>Wattage, it turns out, is critical. Nvidia recommends a system power of 1,000 W for a single RTX 5090, so even a dual-GPU setup exceeds the continuous power draw for a typical 15-ampere, 120-volt power circuit. A system with four RTX 5090s could require 4,000 W or more at load.</p><p><span>The QuietBox 2, on the other hand, draws only 1,400 W at full load. It won’t trip the breaker, so it can be used anywhere a typical desktop PC might be plugged in, including a home office.</span></p><p>That’s not the only way the QuietBox 2 poses as a run-of-the-mill PC. The machine’s custom case is built to support the micro-ATX motherboard form factor, and the motherboard itself is an AMD chipset hosting an AMD CPU. The hardware is kept cool by closed-loop liquid cooling similar to that used by PC workstations and gaming computers. It even has customizable RGB LED lighting and a large semitransparent window that shows off the hardware. </p><p>“A lot of even our internal developers have requested a QuietBox because they’re just so easy to deploy,” says <a href="https://www.linkedin.com/in/chris-goulet-9ab67372/?originalSubdomain=ca" target="_blank">Chris Goulet</a>, a thermal-mechanical engineer and team lead at Tenstorrent. “You just ship them the unit, they slap it on their desk, power it up, and they’re going.” </p><p>Where the QuietBox 2 differs from desktop PCs, though, is its AI accelerators. It’s equipped with four of Tenstorrent’s Blackhole <strong></strong>application-specific ICs, a RISC-V chip designed specifically for AI workloads. Blackhole is packaged on an add-in card; each card has 120 Tensix AI accelerators and 32 GB of GDDR6 memory, for a total of 480 Tensix AI accelerators and 128 GB of GDDR6. Blackhole also has a large amount of on-chip SRAM at 180 megabytes per accelerator.</p><h2>Two visions of desktop AI</h2><p>Tenstorrent is not alone in its approach. Nvidia’s DGX Spark, released last year, packed Nvidia’s GB10 chip into a machine the size of a lunch box. Orders for the Spark’s big brother, the <a href="https://www.nvidia.com/en-us/products/workstations/dgx-station/" target="_blank">DGX Station</a> with Nvidia’s GB300, were opened on 16 March 2026. </p><p>The DGX Station looks like a desktop PC workstation, and variants will be built by well-known PC brands like Asus and Dell. Nvidia’s offering has more memory than QuietBox 2, at up to 748 GB, but system power is quoted at 1,600 W—rather close to the maximum a 15-A, 120-V breaker will handle. This reflects differing visions for how their machines will be used. And, of course, the Nvidia DGX Station’s extra memory doesn’t come cheap. While most DGX Station system builders have not yet announced pricing, <a href="https://www.centralcomputer.com/msi-ct60-s8060-nvidia-dgx-station-cpu-memory-up-to-496gb-lpddr5x-nvidia-blackwell-ultra-gpu-1x-10-gbe-2x-400-gbe.html">one retailer has listed a DGX Station from PC maker MSI for $85,000</a>. </p><p>When I spoke to <a href="https://www.linkedin.com/in/allenbourgoyne/" target="_blank">Allyn Bourgoyne</a>, director of product marketing at Nvidia, after the announcement of DGX Spark and Station in 2025, he said the company expects most DGX owners will use the devices as remotely accessed workstations. “A common thing you might see is that I’ve got my Windows laptop, and I’m going to use my DGX Spark over the network. I’m going to send jobs over to it.” He added that companies could deploy DGX Spark and Station systems to serve multiple people at once. <strong></strong></p><p>The Tenstorrent QuietBox 2 can be used in this way, but the company also wants to target a good experience for people going one-on-one with the computer. “You don’t have to remote SSH into the box. You connect your monitor through HDMI, and it’s just like your PC at home. It has the Ubuntu desktop and utilities,” says Trajkovic.</p><p>Nvidia’s DGX systems also run a variant of Ubuntu (DGX OS) and include a desktop environment, <a href="https://www.jeffgeerling.com/blog/2025/dells-version-dgx-spark-fixes-pain-points/" target="_blank">but the devil is in the details</a>. DGX systems use Nvidia CPUs based on ARM architectures and custom chipsets. The QuietBox 2 uses an AMD x86 CPU and compatible chipset, and is configured more like a traditional PC. That should be a boon for the QuietBox 2’s software compatibility. </p><p>Tenstorrent leans into that with a focus on open source software. The QuietBox 2’s entire software stack, from TT-Forge (the company’s AI compiler) to TT-Metalium (a low-level software development kit that provides kernel-level hardware control), is open source and available on GitHub. Tenstorrent has also published the instruction set architecture for its Tensix cores, so developers can see exactly how their workloads execute on the hardware. Nvidia, by contrast, is focused on its proprietary CUDA ecosystem, and DGX OS is not open source. </p><p>“A lot of our software stack is completely open, and we felt that from a hardware perspective, we kind of wanted to take a similar path,” says Goulet. </p>]]></description><pubDate>Tue, 24 Mar 2026 14:00:05 +0000</pubDate><guid>https://spectrum.ieee.org/ai-workstation-looks-like-pcs</guid><category>Nvidia</category><category>Ai-workstations</category><category>Large-language-models</category><category>Home-computers</category><dc:creator>Matthew S. Smith</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-sleek-illuminated-box-resembling-a-personal-computer-workstation.jpg?id=65329122&amp;width=980"></media:content></item><item><title>The Coming Drone-War Inflection in Ukraine</title><link>https://spectrum.ieee.org/autonomous-drone-warfare</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/person-holding-a-large-drone-outdoors-under-a-sunny-partly-cloudy-sky.jpg?id=65327386&width=1200&height=800&coordinates=0%2C0%2C0%2C0"/><br/><br/><p><strong>WHEN</strong><strong> </strong><strong>KYIV-BORN</strong><strong> </strong><strong>ENGINEER </strong><a href="https://www.instagram.com/yaroslavazhnyuk/?hl=en" rel="noopener noreferrer" target="_blank">Yaroslav Azhnyuk</a> thinks about the future, his mind conjures up dystopian images. He talks about “swarms of autonomous drones carrying other autonomous drones to protect them against autonomous drones, which are trying to intercept them, controlled by <a href="https://spectrum.ieee.org/ai-agents" target="_self">AI</a> <a href="https://spectrum.ieee.org/ai-agents" target="_self">agents</a> overseen by a human general somewhere.” He also imagines flotillas of autonomous submarines, each carrying hundreds of drones, suddenly emerging off the coast of California or Great Britain and discharging their cargoes en masse to the sky.</p><p>“How do you protect from that?” he asks as we speak in late December 2025; me at my quiet home office in London, he in Kyiv, which is bracing for another wave of <a href="https://spectrum.ieee.org/ukraine-air-defense" target="_self">missile attacks</a>.</p><p>Azhnyuk is not an alarmist. He cofounded and was formerly CEO of <a href="https://petcube.com/" rel="noopener noreferrer" target="_blank">Petcube</a>, a California-based company that uses smart cameras and an app to let pet owners keep an eye on their beloved creatures left alone at home. A self-described “liberal guy who didn’t even receive military training,” Azhnyuk changed his mind about developing military tech in the months following the <a href="https://commonslibrary.parliament.uk/research-briefings/cbp-9847/" rel="noopener noreferrer" target="_blank">Russian invasion of</a> <a href="https://commonslibrary.parliament.uk/research-briefings/cbp-9847/" rel="noopener noreferrer" target="_blank">Ukraine</a> in February 2022. By 2023, he had relinquished his CEO role at Petcube to do what many Ukrainian technologists have done—to help defend his country against a mightier aggressor.</p><p>It took a while for him to figure out what, exactly, he should be doing. He didn’t join the military, but through friends on the front line, he witnessed how, out of desperation, Ukrainian troops turned to off-the-shelf consumer drones to make up for their country’s lack of artillery.</p><p>Ukrainian troops first began using drones for battlefield surveillance, but within a few months they figured out how to strap explosives onto them and turn them into effective, <a href="https://spectrum.ieee.org/ukraine-hackers-war" target="_self">low-cost killing</a> <a href="https://spectrum.ieee.org/ukraine-hackers-war" target="_self">machines</a>. Little did they know they were fomenting a revolution in warfare.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Group observes a drone demonstration indoors, with a presenter explaining features." class="rm-shortcode" data-rm-shortcode-id="bfc4f902e7ae9ffa663bf3bcc8ff144c" data-rm-shortcode-name="rebelmouse-image" id="cc3bb" loading="lazy" src="https://spectrum.ieee.org/media-library/group-observes-a-drone-demonstration-indoors-with-a-presenter-explaining-features.jpg?id=65341730&width=980"/></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Compact black camera module with textured surface and orange ribbon cable on white background." class="rm-shortcode" data-rm-shortcode-id="e904e39e8ac7797c354a205ed343d150" data-rm-shortcode-name="rebelmouse-image" id="4d58e" loading="lazy" src="https://spectrum.ieee.org/media-library/compact-black-camera-module-with-textured-surface-and-orange-ribbon-cable-on-white-background.jpg?id=65341726&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">The Ukrainian robotics company The Fourth Law produces an autonomy module [above] that uses optics and AI to guide a drone to its target. Yaroslav Azhnyuk [top, in light shirt], founder and CEO of The Fourth Law, describes a developmental drone with autonomous capabilities to Ukrainian President Volodymyr Zelenskyy and German Chancellor Olaf Scholz.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Top: THE PRESIDENTIAL OFFICE OF UKRAINE; Bottom: THE FOURTH LAW</small></p><p>That revolution was on display last month, as the U.S. and Israel went to war with Iran. It soon became clear that attack drones are being extensively used by both sides. Iran, for example, is relying heavily on the Shahed drones that the country invented and that are now also being manufactured in Russia and launched by the thousands every month against Ukraine.</p><p>A thorough analysis of the Middle East conflict <span>will take some time to emerge. And so to understand the direction of this new way of war, look to Ukraine, where its next phase—autonomy—is already starting to come into view. Outnumbered by the Russians and facing increasingly sophisticated jamming and spoofing aimed at causing the drones to veer off course or fall out of the sky, Ukrainian technologists realized as early as 2023 that what could really win the war was autonomy. Autonomous operation means a drone isn’t being flown by a remote pilot, and therefore there’s no communications link to that pilot that can be severed or spoofed, rendering the drone useless.</span></p><p>By late 2023, <a href="https://www.linkedin.com/in/yaroslavazhnyuk/?locale=uk" target="_blank">Azhnyuk</a> set out to help make that vision a reality. He founded two companies, <a href="https://thefourthlaw.ai/blog/funding-products-video" target="_blank">The</a> <a href="https://thefourthlaw.ai/blog/funding-products-video" target="_blank">Fourth Law</a> and <a href="https://oddsystems.io/en/" target="_blank">Odd Systems</a>, the first to develop AI algorithms to help drones overcome jamming during final approach, the second to build thermal cameras to help those drones better sense their <span>surroundings.</span></p><p>“I moved from making devices that throw treats to dogs to making devices that throw explosives on Russian occupants,” Azhnyuk quips.</p><p>Since then, The Fourth Law has dispatched “more than thousands” of <a href="https://thefourthlaw.ai/#section3" target="_blank">autonomy modules</a> to troops in eastern Ukraine (it declines to give a more specific figure), which can be retrofitted on existing drones to take over navigation during the final <span>approach to the target. Azhnyuk says the autonomy modules, worth around US $50, increase the drone-strike success rate by up to four times that of purely operator-controlled drones.</span></p><p>And that is just the beginning. Azhnyuk is one of thousands of developers, including some <span>who </span>relocated from Western countries, who are applying their skills and other resources to advancing the drone technology that is the defining characteristic <span>of the war in Ukraine. This eclectic group of startups and founders includes </span><a href="https://en.wikipedia.org/wiki/Eric_Schmidt" target="_blank">Eric Schmidt</a>, the forme<a href="https://about.google/company-info/" target="_blank">r</a> <a href="https://about.google/company-info/" target="_blank">Google</a> CEO, whose company <a href="https://epravda.com.ua/oborona/milyarder-ta-ekskerivnik-google-robit-droni-dlya-ukrajini-shcho-nim-ruhaye-809495/" target="_blank">Swift Beat</a> is churning out autonomous <a href="https://www.nytimes.com/2025/12/31/magazine/ukraine-ai-drones-war-russia.html" target="_blank">drones and modules for Ukrainian</a> <a href="https://www.nytimes.com/2025/12/31/magazine/ukraine-ai-drones-war-russia.html" target="_blank">forces</a>. The frenetic pace of tech development is helping a scrappy, innovative underdog hold at bay a much larger and better-equipped foe.</p><p>All of this development is careening toward AI-based systems that enable drones to navigate by recognizing features in the terrain, lock on to and chase targets without an operator’s guidance, and eventually exchange information with each other through mesh networks, forming self-organizing robotic kamikaze swarms. Such an attack swarm would be commanded by a single operator from a safe distance.</p><p><span>According to some reports, autonomous swarming technology is also being developed <a href="https://www.usni.org/magazines/proceedings/2025/may/step-step-ukraine-built-technological-navy" target="_blank">for</a> <a href="https://www.usni.org/magazines/proceedings/2025/may/step-step-ukraine-built-technological-navy" target="_blank">sea drones</a>. Ukraine has had some notable <span>successes with sea drones, which have reportedly</span> </span><span>destroyed or damaged </span><a href="https://en.usm.media/sbu-naval-drones-hit-11-russian-ships-and-vessels-details/" target="_blank">around a dozen</a><span> Russian vessels.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Hand holding a drone with six rotors, outdoors against a blue sky." class="rm-shortcode" data-rm-shortcode-id="90f30978c5ba0e77e9b1873c155131d2" data-rm-shortcode-name="rebelmouse-image" id="7cf11" loading="lazy" src="https://spectrum.ieee.org/media-library/hand-holding-a-drone-with-six-rotors-outdoors-against-a-blue-sky.jpg?id=65341722&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">The Skynode X system, from Auterion, provides a degree of autonomy to a drone.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">AUTERION</small></p><p>For Ukraine, swarming can solve a major problem that puts the nation at a disadvantage against Russia—the lack of personnel. Autonomy is “the single most impactful defense technology of this century,” says Azhnyuk. “The moment this happens, you <span>shift from a manpower challenge to a production challenge, which is much more manageable,” he adds.</span></p><p>The autonomous warfare future envisioned by Azhnyuk and others is not yet a reality. But <a href="https://www.linkedin.com/in/marcclange/?skipRedirect=true" target="_blank">Marc Lange</a>, a German defense analyst and business strategist, believes that “an inflection point” is already in view. Beyond it, “things will be so dramatically different,” he says.</p><p>“Ukraine pretty rapidly realized that if the operator-to-drone ratio can be shifted from one-to-one to one-to-many, that creates great economies of scale and an amazing cost exchange ratio,” Lange adds. “The moment one operator can launch 100, 50, or even just 20 drones at once, this completely changes the economics of the war.”</p><h2>Drones With a View </h2><p>For a while, jammers that sever the radio links between drones and <span>operators or that spoof GPS receivers were able to provide fairly reliable defense against human-controlled first-person-view attack drones (FPVs). But as autonomous navigation progressed, those electronic shields have gradually become less effective. Defenders must now contend with unjammable drones—ones that are attached to hair-thin optical fibers or that are capable of </span><a href="https://spectrum.ieee.org/ukraine-killer-drones" target="_self">finding</a> <a href="https://spectrum.ieee.org/ukraine-killer-drones" target="_self">their way to their targets</a> without external guidance. In this emerging struggle, the defenders’ track records aren’t very encouraging: The typical countermeasure is to try to shoot down the attacking drone with a service weapon. It’s rarely successful.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Truck on rural road covered with camouflage netting, trees and fields in the background." class="rm-shortcode" data-rm-shortcode-id="7c7af1e395cf35752b367f8dd54130fc" data-rm-shortcode-name="rebelmouse-image" id="58155" loading="lazy" src="https://spectrum.ieee.org/media-library/truck-on-rural-road-covered-with-camouflage-netting-trees-and-fields-in-the-background.jpg?id=65341708&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">A truck outfitted with signal-jamming gear drives under antidrone nets near Oleksandriya, in eastern Ukraine, on 2 October 2025.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">ED JONES/AFP/GETTY IMAGES</small></p><p>“The attackers gain an immense advantage from unmanned systems,” says Lange. “You can have a drone pop up from anywhere and it can wreak havoc. But from autonomy, they gain even more.”</p><p>The self-navigating drones rely on image-recognition algorithms that have been around for over a decade, says Lange. And the mass deployments of drones on Ukrainian battlefields are enabling both Russian and Ukrainian technologists to create <a href="https://www.reuters.com/technology/ukraine-collects-vast-war-data-trove-train-ai-models-2024-12-20/" target="_blank">huge datasets</a> that improve the training and precision of those AI algorithms.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Six-wheeled robotic vehicle with mounted equipment in a grassy field." class="rm-shortcode" data-rm-shortcode-id="caa0a697b2d5752603687ac7f0278581" data-rm-shortcode-name="rebelmouse-image" id="1c591" loading="lazy" src="https://spectrum.ieee.org/media-library/six-wheeled-robotic-vehicle-with-mounted-equipment-in-a-grassy-field.jpg?id=65341706&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">A Ukrainian land robot, the Ravlyk, can be outfitted with a machine gun.</small></p><p>While uncrewed aerial vehicles (UAVs) have received the most attention, the Ukrainian military is also deploying dozens of different kinds of drones on land and sea. Ukraine, struggling with the shortage of infantry personnel, began working on replacing a portion of human soldiers with wheeled ground robots in 2024. As of early 2026, thousands of ground robots are crawling across the gray zone along the front line in Eastern Ukraine. Most are used to deliver supplies to the front line or to help evacuate the wounded, but some “killer” ground robots fitted with turrets and remotely controlled machine guns have also been tested.</p><p>In mid-February, Ukrainian authorities released a video of a Ukrainian ground robot using its thermal camera to detect a Russian soldier in the dark of the night and then kill the invader with a round from a heavy machine gun. So far these robots are mostly controlled <span>by a human operator, but the makers of these uncrewed ground vehicles say their systems are capable of basic autonomous operations, such as returning to base when radio connection is lost. The goal is to enable them to swarm so that one operator controls not one, but a whole herd of mesh-connected killer robots.</span></p><p>But <a href="https://www.hudson.org/experts/1303-bryan-clark" target="_blank">Bryan <span>Clark</span></a>, senior fellow and <span>director of the Center for Defense Concepts and Technology at the </span><a href="https://www.hudson.org/" target="_blank">Hudson Institute</a>, questions how quickly ground robots’ abilities can progress. “Ground environments are very difficult to navigate in because of the terrain you have to address,” he says. “The line of sight for the sensors on the ground vehicles is really constrained because of terrain, whereas an air vehicle can see everything around it.”</p><p>To achieve autonomy, <a href="https://spectrum.ieee.org/sea-drone" target="_self">maritime drones</a>, too, will require <span>naviga</span><span>tional approaches beyond AI-based image recognition, possibly based on star positions or electronic signals from radios and cell towers that are within reach, says Clark. Such technologies are still being developed or are in a relatively early operational stage.</span></p><h2>How the Shaheds Got Better</h2><p>Russia is not lagging behind. In fact, some analysts believe its autonomous systems may be slightly ahead of Ukraine’s. For a good example of the Russian military’s rapid <span>evolu</span><span>tion, they say, consider the long-range Iranian-designed Shahed drones. Since 2022, Russia has been using them to attack Ukrainian cities and other targets hundreds of kilometers from the front line. “At the beginning, Shaheds just had a frame, a </span><span>motor, and an inertial navigation system,” </span><a href="https://www.linkedin.com/in/oleksii-solntsev-aa0b72189?originalSubdomain=ua" target="_blank">Oleksii</a><span> </span><a href="https://www.linkedin.com/in/oleksii-solntsev-aa0b72189?originalSubdomain=ua" target="_blank">Solntsev</a><span>, CEO of Ukrainian defense tech startup MaXon Systems, tells me. “They used to be imprecise and pretty stupid. But they are becoming more and more autonomous.” Solntsev founded MaXon </span><span>Systems in late 2024 to help protect Ukrainian civil</span><span>ians from the growing threat of Shahed </span><span>raids.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Silhouette of a triangular drone flying in the sky." class="rm-shortcode" data-rm-shortcode-id="a9c89e21028ccf85e20a49ecead8309f" data-rm-shortcode-name="rebelmouse-image" id="72159" loading="lazy" src="https://spectrum.ieee.org/media-library/silhouette-of-a-triangular-drone-flying-in-the-sky.jpg?id=65341701&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">A Russian Geran-2 drone, based on the Iranian Shahed-136, flies over Kyiv during an attack on 27 December 2025.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">SERGEI SUPINSKY/AFP/GETTY IMAGES</small></p><p>First produced <a href="https://www.adaptinstitute.org/from-tehran-to-alabuga-the-evolution-of-shahed-drones-into-russias-strategic-asset/26/09/2025/" target="_blank">in Iran in the 2010s</a>, Shaheds can <span>carry 90-kilogram warheads </span><a href="https://isis-online.org/isis-reports/alabugas-shahed-136-geran-2-warheads-a-dangerous-escalation" target="_blank">up to 650 km</a> (50-kg warheads can go twice as far). <a href="https://www.csis.org/analysis/calculating-cost-effectiveness-russias-drone-strikes" target="_blank">They cost around $35,000 per unit</a><span>, compared to a couple of million dollars, at least, for a ballistic missile. The low cost </span><span>allows Russia to manufacture Shaheds in high quantities, unleashing entire fleets onto </span><a href="https://isis-online.org/isis-reports/a-comprehensive-analytical-review-of-russian-shahed-type-uavs-deployment-against-ukraine-in-2025" target="_blank">Ukrainian cities</a><span> </span><a href="https://isis-online.org/isis-reports/a-comprehensive-analytical-review-of-russian-shahed-type-uavs-deployment-against-ukraine-in-2025" target="_blank">and infrastructure almost every night</a><span>.</span></p><p>The early Shaheds were able to reach a prepro<span>grammed location based on satellite-navigation coordinates. Even one of these early models could frequently overcome the jamming of satellite-navigation signals with the help of an onboard inertial navigation unit. This was essentially a dead-reckoning system of accelerators and gyroscopes that estimate the drone’s position from continual measurements of its motions.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Silhouette of person with large equipment under a starry night sky." class="rm-shortcode" data-rm-shortcode-id="37186ec06b71203ba4f30db497507797" data-rm-shortcode-name="rebelmouse-image" id="1aca7" loading="lazy" src="https://spectrum.ieee.org/media-library/silhouette-of-person-with-large-equipment-under-a-starry-night-sky.jpg?id=65341699&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">In the Donetsk Region, on 15 August 2025, a Ukrainian soldier hunts for Shaheds and other drones with a thermalimaging system attached to a ZU23 23-millimeter antiaircraft gun.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">KOSTYANTYN LIBEROV/LIBKOS/GETTY IMAGES</small></p><p>Ukrainian defense forces learned to down Shaheds with heavy machine guns, but as Russia continued to innovate, the daily onslaughts started to become <a href="https://euromaidanpress.com/2025/06/29/why-cant-ukraine-stop-russias-shahed-drones-anymore/" target="_blank">increasingly effective.</a></p><p>Today’s Shaheds fly faster and higher, and therefore are more difficult to detect and take down. Between January 2024 and August 2025, the number of Shaheds and Shahed-type attack drones launched by Russia into Ukraine per month <a href="https://united24media.com/war-in-ukraine/why-russias-shahed-drones-are-now-deadlier-and-harder-than-ever-to-stop-11693" target="_blank">increased more than tenfold</a>, from 334 to more than 4,000. In 2025, Ukraine found <a href="https://www.unmannedairspace.info/counter-uas-systems-and-policies/recently-downed-russian-shahed-demonstrates-new-levels-of-autonomous-capability/" target="_blank">AI-enabling</a> <a href="https://www.unmannedairspace.info/counter-uas-systems-and-policies/recently-downed-russian-shahed-demonstrates-new-levels-of-autonomous-capability/" target="_blank">N</a><a href="https://www.unmannedairspace.info/counter-uas-systems-and-policies/recently-downed-russian-shahed-demonstrates-new-levels-of-autonomous-capability/" target="_blank">vidia</a> <a href="https://www.unmannedairspace.info/counter-uas-systems-and-policies/recently-downed-russian-shahed-demonstrates-new-levels-of-autonomous-capability/" target="_blank">chipsets in wreckages of Shaheds</a>, as well as thermal-vision modules capable of locking onto targets at night.</p><p>“Now, they are interconnected, which allows them to exchange information with each other,” Solntsev says. “They also have cameras that allow them to autonomously navigate to objects. Soon they will be able to tell each other to avoid a <span>jammed</span> <span>region or an area where one of them got </span><span>intercepted.”</span></p><p>These Russian-manufactured Shaheds, which Russian forces call Geran-2s, are thought to be more capable than the garden variety Shahed-136s that Iran has lately been launching against targets throughout the Middle East. Even the relatively primitive Shahed-136s have done considerable damage, according to <a href="https://www.theguardian.com/world/2026/mar/02/iran-unleashes-hundreds-of-drones-aimed-at-targets-across-middle-east" target="_blank">press accounts</a>.</p><p>Those Shahed successes may accrue, at least in part, from the fact that the United States and Israel <span>lack Ukraine’s long experience with fending them off. In just two days in early March, upward of a thousand drones, mostly Shaheds, were launched against U.S. and Israeli targets, with </span><a href="https://www.theguardian.com/world/2026/mar/02/iran-unleashes-hundreds-of-drones-aimed-at-targets-across-middle-east" target="_blank">hundreds of</a> <a href="https://www.theguardian.com/world/2026/mar/02/iran-unleashes-hundreds-of-drones-aimed-at-targets-across-middle-east" target="_blank">them reportedly finding their marks</a>.</p><p>One attack, caught on videotape, shows a Shahed destroying a radar dome at the U.S. navy base in <span>Manama, Bahrain. U.S. forces were understood to be </span><a href="https://carnegieendowment.org/emissary/2026/03/iran-drones-shahed-us-lessons" target="_blank">attempting to fend off the drones</a> by striking launch platforms, dispatching fighter aircraft to shoot them down, and by using some extremely costly air-defense interceptors, including ones meant to down ballistic missiles. On 4 March, <a href="https://www.cnn.com/2026/03/04/politics/us-air-defenses-iran-attack-drones-challenge" target="_blank">CNN</a> <a href="https://www.cnn.com/2026/03/04/politics/us-air-defenses-iran-attack-drones-challenge" target="_blank">reported</a> that in a congressional briefing the day before, top U.S. defense officials, including Secretary of Defense Pete Hegseth, acknowledged that U.S. air defenses weren’t keeping up with the onslaught of Shahed drones.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Broken drone on soil, cylindrical container nearby." class="rm-shortcode" data-rm-shortcode-id="769830682ff53a401780108ca11db2b6" data-rm-shortcode-name="rebelmouse-image" id="c9d58" loading="lazy" src="https://spectrum.ieee.org/media-library/broken-drone-on-soil-cylindrical-container-nearby.jpg?id=65341692&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">Russian V2U attack drones are outfitted with Nvidia processors and run computer-vision software and AI algorithms to enable the drones to navigate autonomously.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">GUR OF THE MINISTRY OF DEFENSE OF UKRAINE</small></p><p>Russia is also starting to field a newer generation of attack drones. One of these, the V2U, has been used to strike targets in the Sumy region of northeastern Ukraine. <a href="https://euromaidanpress.com/2025/06/09/russias-v2u-drone-uses-ai-for-autonomous-strikes-in-ukraines-sumy-oblast/" target="_blank"><span>The V2U drones</span></a> are outfitted with Nvidia Jetson Orin processors and run <span>computer</span>-­<span>vision software and AI algorithms that allow the drones to navigate even where satellite navigation is jammed.</span></p><p>The sale of Nvidia chips to Russia is banned under U.S. sanctions against the country. However, press reports suggest that the chips are getting to Russia <a href="https://www.pravda.com.ua/eng/news/2024/10/28/7481703/" target="_blank">via intermediaries in India</a>.</p><h2>Antidrone Systems Step Up</h2><p>MaXon Systems is one of several companies working to fend off the nightly drone onslaught. Within one year, the company developed and battle-tested a Shahed interception system that hints at the sci-fi future envisioned by Azhnyuk. For a system to be capable of reliably defending against autonomous weaponry, it, too, needs to be autonomous.</p><p><span>MaXon’s solution consists of ground turrets scanning the sky with infrared sensors, with additional input from a network of radars that </span><span>detects approaching Shahed drones at distances of, typically, </span><a href="https://en.defence-ua.com/weapon_and_tech/2025_systems_to_shield_kyiv_from_shaheds_new_air_defense_details_from_maxon_where_balloons_carry_interceptor_drones-15499.html" target="_blank">12 to 16</a><span> km. The turrets fire autonomous fixed-winged interceptor drones, fitted with explosive warheads, toward the approaching Shaheds at speeds of nearly 300 km/h. To boost the chances of successful interception, MaXon </span><a href="https://en.defence-ua.com/weapon_and_tech/2025_systems_to_shield_kyiv_from_shaheds_new_air_defense_details_from_maxon_where_balloons_carry_interceptor_drones-15499.html" target="_blank">is also fielding</a><span> an airborne anti-Shahed fortification </span><span>system</span><span> </span><span>consisting of helium-filled </span><a href="https://spectrum.ieee.org/airships-drones-ukraine" target="_self">aerostats</a><span> hovering above the city that dispatch the interceptors from a higher altitude.</span></p><p>“We are trying to increase the level of automation of the system compared to existing solutions,” says Solntsev. “We need automatic <span>detection, automatic takeoff, and automatic mid-track guidance so that we can guide the interceptor before it can itself flock the target.”</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Gray drone on display stand, surrounded by military personnel in camouflage uniforms." class="rm-shortcode" data-rm-shortcode-id="592b19dbfc4fe9a54033067c6169aeec" data-rm-shortcode-name="rebelmouse-image" id="ab79b" loading="lazy" src="https://spectrum.ieee.org/media-library/gray-drone-on-display-stand-surrounded-by-military-personnel-in-camouflage-uniforms.jpg?id=65341687&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">An interceptor drone, part of the U.S. MEROPS defensive system, is tested in Poland on 18 November 2025.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">WOJTEK RADWANSKI/AFP/GETTY IMAGES</small></p><p>In November 2025, the Ukrainian military announced it had been conducting successful trials of the <a href="https://www.forcesnews.com/nato/bang-your-buck-200m-worth-russian-drones-taken-out-15m-merops-uavs" target="_blank">Merops Shahed drone interceptor</a> system developed by the U.S. startup <a href="https://themerge.co/p/project-eagle" target="_blank">Project Eagle</a>, another of former <span>Google CEO Eric Schmidt’s Ukraine defense ventures. Like the MaXon gear, the system can operate largely autonomously and has so far downed over 1,000 Shaheds.</span></p><h2>What Works in the Lab Doesn’t Necessarily Fly on the Battlefield </h2>Despite the progress on both sides, analysts say that <span>the kind of robotic warfare imagined by Azhnyuk won’t be a reality for years.</span><p>“The software for drone collaboration is there,” says <a href="https://www.csis.org/people/kateryna-bondar" target="_blank">Kate Bondar</a>, a former policy advisor for the Ukrainian <span>government and currently a research fellow at the U.S. </span><a href="https://www.csis.org/" target="_blank">Center for Stra</a><a href="https://www.csis.org/" target="_blank">tegic and International Studies</a><span>. “Drones can fly in labs, but in real life, [the forces] are afraid to deploy them because the risk of a mistake is too high,” she adds.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Two people launching a drone in an open field using a catapult system." class="rm-shortcode" data-rm-shortcode-id="894baf9e936bef6f8c45a0363afac141" data-rm-shortcode-name="rebelmouse-image" id="7c4e9" loading="lazy" src="https://spectrum.ieee.org/media-library/two-people-launching-a-drone-in-an-open-field-using-a-catapult-system.jpg?id=65341682&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">Ukrainian soldiers watch a GOR reconnaissance drone take to the sky near Pokrovsk in the Donetsk region, on 10 March 2025.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">ANDRIY DUBCHAK/FRONTLINER/GETTY IMAGES</small></p>In Bondar’s view, powerful AI-equipped drones won’t be deployed in large numbers given the current prices for high-end processors and <span>other advanced components. And, she adds, the more autonomous the system needs to be, the more expensive are the processors and sensors it must have. “For these cheap attack drones that fly only once, you don’t install a high-resolution camera that [has] the resolution for AI to see properly,” she says. “[You install] the cheapest camera. You don’t </span><span>want expensive chips that can run AI algorithms either. Until we can achieve this balance of technological sophistication, when a system can conduct a mission but at the lowest price possible, it won’t be deployed en masse.”</span><p>While existing AI systems are doing a good job recognizing and following large objects like Shaheds or tanks, experts question their ability to reliably distinguish and pursue smaller and more nimble or inconspicuous targets. “When we’re getting into more specific questions, like can it distinguish a Russian soldier from a Ukrainian soldier or at least a soldier from a civilian? The answer is no,” says Bondar. “Also, it’s one thing to track a tank, and it’s another to track infantrymen riding buggies and motorcycles that are moving very fast. That’s really challenging for AI to track and strike precisely.”</p><p>Clark, at the Hudson Institute, says that although the AI algorithms used to guide the Russian and <span>Ukrainian drones are “pretty good,” they rely on information provided bysensors that “aren’t good enough.” “You need multiphenomenology sensors that are able to look at infrared and visual and, in some cases, different parts of the infrared spectrum to be able to figure out if something is a decoy or real target,” </span><span>he </span><span>says.</span></p><p><span>German defense analyst Lange agrees that right now, battlefield AI image-recognition systems are too easily fooled. “If you compress reality into a </span><span>2D</span><span> image, a lot of things can be easily camouflaged—like what Russia did recently, when they started drawing birds on the back of their drones,” he <span>says.</span></span></p><h2>Autonomy Remains Elusive on the Ground and at Sea, Too</h2><p>To make Ukraine’s <span>emerging uncrewed ground vehicles (UGVs) equally self-sufficient will be an even greater task, in Clark’s view. Still, </span><span>Bondar expects major advances to materialize within the next several years, even if humans are still going to be part of the decision-making loop.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Military radar equipment in a grassy field." class="rm-shortcode" data-rm-shortcode-id="0b36a03b7582535b3d3319d7d9b74c33" data-rm-shortcode-name="rebelmouse-image" id="d65ea" loading="lazy" src="https://spectrum.ieee.org/media-library/military-radar-equipment-in-a-grassy-field.jpg?id=65341671&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">A mobile electronic-warfare system built by PiranhaTech is demonstrated near Kyiv on 21 October 2025.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">DANYLO ANTONIUK/ANADOLU/GETTY IMAGES</small></p><p>“I think in two or three years, we will have pretty good full autonomy, at least in good weather conditions,” she says, referring to aerial drones in partic<span>ular. “Humans will still be in the loop for some years, simply because there are so many unpredictable situations when you need an intervention. We won’t be able to fully rely on the machine for at least another 10 or 15 years.”</span></p><p>Ukrainian defenders are apprehensive about that autonomous future. The boom of drone inno<span>vation has come hand in hand with the development of sophisticated jamming and radio-frequency detection systems. But a lot of that innovation will become obsolete once the pendulum swings away from human control. Ukrainians got their first taste of dealing with unjammable drones in mid-2024, when Russia began rolling out fiber-optic tethered drones. Now they have to brace for a threat on a much larger scale.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Quadcopter drone flying with a fire extinguisher attached in a cloudy sky." class="rm-shortcode" data-rm-shortcode-id="70f326221988cb6004338272d1d8dd4d" data-rm-shortcode-name="rebelmouse-image" id="aa25d" loading="lazy" src="https://spectrum.ieee.org/media-library/quadcopter-drone-flying-with-a-fire-extinguisher-attached-in-a-cloudy-sky.jpg?id=65341673&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">An experimental drone is demonstrated at the Brave1 defense-tech incubator in Kyiv.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">DANYLO DUBCHAK/FRONTLINER/GETTY IMAGES</small></p><p>“Today, we have a situation where we have lots of signals on the battlefield, but in the near future, <span>in maybe two to five years, UAVs are not going to be sending any signals,” says Oleksandr Barabash, CTO of </span><a href="https://www.falcons.com.ua/en" target="_blank">Falcons</a>, a Ukrainian startup that has developed a smart radio-frequency detection system capable <span>of revealing precise locations of enemy radio sources such as drones, control stations, and jammers.</span></p><p>Last September, Falcons secured funding from the U.S.-based dual-use tech fund <a href="https://www.greenflag.vc/" target="_blank">Green Flag Ven</a><a href="https://www.greenflag.vc/" target="_blank">tures</a> to scale production of its technology and work toward NATO certification. But Barabash admits that its system, like all technologies fielded in <span>Ukrainian war zones, has an expiration date. Instead of radio-frequency detectors, Barabash thinks, the next R&D push needs to focus on passive radar systems capable of identifying small and fast-moving targets based on the signal from sources like TV towers or radio transmitters that propagate through the environment and are reflected by those moving targets. Passive radars have a significant advantage in the war zone, according to Barabash. Since they don’t emit their own signal, they can’t be that easily discovered by the enemy.</span></p><p>“Active radar is emitting signals, so if you are using active radars, you are target No. 1 on the front line,” Barabash says.</p><p><span>Bondar, on the other hand, thinks that the increased onboard compute power needed </span><span>for</span> AI-controlled drones will, by itself, generate enough electromagnetic radiation to prevent autonomous drones from ever operating completely undetectably.</p><p><span>“You can have full autonomy, but you will still have systems onboard that emit electromagnetic radiation or heat that can be detected,” says Bondar. “Batteries emit electromagnetic radiation, motors emit heat, and [that heat can be] visible in infrared from far away. You just need to have the right sensors to be able to identify it in advance.” She adds that that takeaway is “how capable contemporary detection systems have become and how technically challenging it is to design drones that can reliably operate in the Ukrainian battlefield environment.”</span></p><h2>There Will Be Nowhere to Hide from Autonomous Drones</h2><p>When autonomous drones become a standard weapon <span>of war, their threat will extend far beyond the battlefields of Ukraine. Autonomous turrets and drone-interceptor fortification might soon dot the perimeter of European cities, particularly in the eastern part of the continent.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Person holding gray drone against a blue sky, preparing to launch it." class="rm-shortcode" data-rm-shortcode-id="c480e8fb2bdf2e560c142729e35c7320" data-rm-shortcode-name="rebelmouse-image" id="f9032" loading="lazy" src="https://spectrum.ieee.org/media-library/person-holding-gray-drone-against-a-blue-sky-preparing-to-launch-it.jpg?id=65327903&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">A fixed-wing drone is tested in Ukraine in April 2025.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">ANDREWKRAVCHENKO/BLOOMBERG/GETTY IMAGES</small></p><p>Nefarious actors from all over the world have closely watched Ukraine and taken notes, warns Lange. Today, FPV drones are being used b<a href="https://gnet-research.org/2025/07/30/weaponised-skies-the-expansion-of-terrorist-drone-use-across-africa/" target="_blank">y</a> <a href="https://gnet-research.org/2025/07/30/weaponised-skies-the-expansion-of-terrorist-drone-use-across-africa/" target="_blank">Islamic terrorists in Africa</a> and <a href="https://www.atlanticcouncil.org/blogs/new-atlanticist/drug-cartels-are-adopting-cutting-edge-drone-technology-heres-how-the-us-must-adapt/#%3A~%3Atext%3DIf%20confirmed%2C%20this%20would%20suggest%2CUS%20homeland%20security%E2%80%94are%20profound" target="_blank">Mexican drug cartels</a> to fight against local authorities.</p><p>When autonomous killing machines become widely available, it’s likely that no city will be safe. “We might see nets above city centers, protecting civilian streets,” Lange says. “In every case, the West needs to start performing similar kinetic-defense development that we see in Ukraine. Very rapid iteration and testing cycles to find solutions.”</p><p>Azhnyuk is concerned that the historic defenders of Europe—the <span>United States and the European countries themselves—are falling behind. “We are in danger,” he says. While Russia and Ukraine made major strides in their drones and countermeasures over the past year, “Europe and the United States have progressed, in the best-case scenario, from the winter-of-2022 technology to the summer-of-2022 technology.</span></p><p>“The gap is getting wider,” he warns. “I think the next few years are very dangerous for the security of Europe.” <span class="ieee-end-mark"></span></p><p><em>This article appears in the April 2026 print issue as “Rise of the <span>AUTONOMOUS </span>Attack Drones.”</em></p>]]></description><pubDate>Tue, 24 Mar 2026 13:00:05 +0000</pubDate><guid>https://spectrum.ieee.org/autonomous-drone-warfare</guid><category>Military-robots</category><category>Military-drones</category><category>Drone-war</category><category>Shahed-drones</category><category>Ai-agents</category><dc:creator>Tereza Pultarova</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/person-holding-a-large-drone-outdoors-under-a-sunny-partly-cloudy-sky.jpg?id=65327386&amp;width=980"></media:content></item><item><title>Transforming Data Science With NVIDIA RTX PRO 6000 Blackwell Workstation Edition</title><link>https://spectrum.ieee.org/nvidia-rtx-pro-6000-pny</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/computer-setup-with-a-monitor-displaying-forest-graphics-keyboard-mouse-and-a-sleek-cpu-design.png?id=65315285&width=1200&height=800&coordinates=19%2C0%2C19%2C0"/><br/><br/><p><em>This is a sponsored article brought to you by <a href="https://www.pny.com/" target="_blank">PNY Technologies</a>.</em></p>In today’s data-driven world, data scientists face mounting challenges in preparing, scaling, and processing massive datasets. Traditional CPU-based systems are no longer sufficient to meet the demands of modern AI and analytics workflows. <a href="https://www.pny.com/nvidia-rtx-pro-6000-blackwell-ws?iscommercial=true&utm_source=IEEE+Spectrum+Blog&utm_medium=RTX+PRO+6000+body&utm_campaign=Blackwell+Workstation&utm_id=RTX+PRO+6000" rel="noopener noreferrer" target="_blank">NVIDIA RTX PRO<sup>TM</sup> 6000 Blackwell Workstation Edition</a> offers a transformative solution, delivering accelerated computing performance and seamless integration into enterprise environments.<h2>Key Challenges for Data Science</h2><ul><li><strong>Data Preparation: </strong>Data preparation is a complex, time-consuming process that takes most of a data scientist’s time.</li><li><strong>Scaling: </strong>Volume of data is growing at a rapid pace. Data scientists may resort to downsampling datasets to make large datasets more manageable, leading to suboptimal results.</li><li><strong>Hardware: </strong>Demand for accelerated AI hardware for data centers and cloud service providers (CSPs) is exceeding supply. Current desktop computing resources may not be suitable for data science workflows.</li></ul><h2>Benefits of RTX PRO-Powered AI Workstations</h2><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition delivers ultimate acceleration for data science and AI workflows. These powerful and robust workstations enable real-time rendering, rapid prototyping, and seamless collaboration. With support for up to four <a href="https://www.pny.com/nvidia-rtx-pro-6000-blackwell-max-q?iscommercial=true&utm_source=IEEE+Spectrum+Blog&utm_medium=RTX+PRO+6000+Blackwell+Max-Q+body&utm_campaign=Blackwell+Workstation&utm_id=RTX+PRO+6000" rel="noopener noreferrer" target="_blank">NVIDIA RTX PRO 6000 Blackwell Max-Q Workstation Edition</a> GPUs, users can achieve data center-level performance right at their desk, making even the most demanding tasks manageable.</p><p class="shortcode-media shortcode-media-youtube"> <span class="rm-shortcode" data-rm-shortcode-id="61bf7564ac8304e10487689487367c94" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/jwxxgHsU1jA?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span> <small class="image-media media-caption" placeholder="Add Photo Caption...">PNY is redefining professional computing with the ‪@NVIDIA‬ RTX PRO 6000 Blackwell Workstation Edition, the most powerful desktop GPU ever built. Engineered for unmatched compute power, massive memory capacity, and breakthrough performance, this cutting-edge solution delivers a quantum leap forward in workflow efficiency, enabling professionals to tackle the most demanding applications with ease.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">PNY</small></p><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition empowers data scientists to handle massive datasets, perform advanced visualizations, and support multi-user environments without compromise. It’s ideal for organizations scaling up their analytics or running complex models. NVIDIA RTX PRO 6000 Blackwell Workstation Edition is optimized for AI workflows, leveraging the NVIDIA AI software stack, including CUDA-X, and NVIDIA Enterprise software. These platforms enable zero-code-change acceleration for Python-based workflows and support over 100 AI-powered applications, streamlining everything from data preparation to model deployment.</p><p>Finally, NVIDIA RTX PRO 6000 Blackwell Workstation Edition offers significant advantages in security and cost control. By offloading compute from the data center and reducing reliance on cloud resources, organizations can lower expenses and keep sensitive data on-premises for enhanced protection.</p><h2>Accelerate Every Step of Your Workflow</h2><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition is designed to transform the entire data science pipeline, delivering end-to-end acceleration from data preparation to model deployment. With NVIDIA CUDA-X open-source data science cuDF library and other GPU-accelerated libraries, data scientists can process massive datasets at lightning speed, often achieving up to 50X faster performance compared to traditional CPU-based tools. This means tasks like cleaning data, managing missing values, and engineering features can be completed in seconds, not hours, allowing teams to focus on extracting insights and building better models.</p><p class="pull-quote">NVIDIA RTX PRO 6000 Blackwell Workstation Edition is designed to transform the entire data science pipeline, delivering end-to-end acceleration from data preparation to model deployment</p><p>Exploratory data analysis is elevated with advanced analytics and interactive visualizations, powered by NVIDIA CUDA-X and PyData libraries. These tools enable users to create expansive, responsive visualizations that enhance understanding and support critical decision-making. When it comes to model training, GPU-accelerated XGBoost slashes training times from weeks to minutes, enabling rapid iteration and faster time-to-market AI solutions.</p><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition streamlines collaboration and scalability. With NVIDIA AI Workbench, teams can set up projects, develop, and collaborate seamlessly across desktops, cloud platforms, and data centers. The unified software stack ensures compatibility and robustness, while enterprise-grade hardware maximizes uptime and reliability for demanding workflows.</p><p>By integrating these advanced capabilities, NVIDIA RTX PRO 6000 Blackwell Workstation Edition empowers data scientists to overcome bottlenecks, boost productivity, and drive innovation, making them an essential foundation for modern, enterprise-ready AI development.</p><h2>Performance Benchmarks</h2><p>NVIDIA’s cuDF library offers zero-code change acceleration for pandas, delivering up to 50X performance gains. For example, a join operation that takes nearly 5 minutes on CPU completes in just 14 seconds on GPU. Advanced group by operations drop from almost 4 minutes to just 4 seconds.</p><h2>Enterprise-Ready Solutions from PNY</h2><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" rel="float: left;" style="float: left;"> <img alt="Black PNY logo with stylized uppercase letters on a transparent background." class="rm-shortcode" data-rm-shortcode-id="247ffcd9e141f1fc61c5172c5440d97e" data-rm-shortcode-name="rebelmouse-image" id="170af" loading="lazy" src="https://spectrum.ieee.org/media-library/black-pny-logo-with-stylized-uppercase-letters-on-a-transparent-background.png?id=65315393&width=980"/></p><p>Available from leading OEM manufacturers, NVIDIA RTX PRO 6000 Blackwell Workstation Edition Series GPUs are specifically engineered to meet the rigorous demands of enterprise environments. These systems incorporate NVIDIA Connect-X networking, now available at PNY and a comprehensive suite of deployment and support tools, ensuring seamless integration with existing IT infrastructure.</p><p>Designed for scalability, the latest generation of workstations can tackle complex AI development workflows at scale for training, development, or inferencing. Enterprise-grade hardware maximizes uptime and reliability.</p><p><strong>To learn more about NVIDIA RTX PRO™ Blackwell solutions, </strong><strong>visit:</strong> <a href="https://www.pny.com/professional/software-solutions/blackwell-architecture?utm_source=IEEE+Spectrum+Blog&utm_medium=Blackwell+Desktop+GPUs+learn+more&utm_campaign=Blackwell+Workstation&utm_id=RTX+PRO+6000" target="_blank">NVIDIA RTX PRO Blackwell | PNY Pro | pny.com</a> or email <a href="mailto:gopny@pny.com" target="_blank">GOPNY@PNY.COM</a><strong></strong></p>]]></description><pubDate>Mon, 23 Mar 2026 13:00:04 +0000</pubDate><guid>https://spectrum.ieee.org/nvidia-rtx-pro-6000-pny</guid><category>Artificial-intelligence</category><category>Computing</category><category>Data-science</category><category>Gpu-acceleration</category><category>Ai-workstations</category><category>Nvidia</category><dc:creator>PNY Technologies</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/computer-setup-with-a-monitor-displaying-forest-graphics-keyboard-mouse-and-a-sleek-cpu-design.png?id=65315285&amp;width=980"></media:content></item><item><title>Why Thermal Metrology Must Evolve for Next-Generation Semiconductors</title><link>https://content.knowledgehub.wiley.com/heat-beneath-the-surface-thermal-metrology-for-advanced-semiconductor-materials-and-architectures/</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/laser-thermal-logo-with-stylized-red-l-and-t-on-a-white-background.png?id=65320713&width=980"/><br/><br/><p>An in-depth examination of how rising power density, 3D integration, and novel materials are outpacing legacy thermal measurement — and what advanced metrology must deliver.</p><p><strong>What Attendees will Learn</strong></p><ol><li><span>Why heat is now the dominant constraint on semiconductor scaling — Explore how heterogeneous integration, 3D stacking, and AI-driven power density have shifted the primary bottleneck from lithography to thermal management, with heat flux projections exceeding 1,000 W/cm² for next-generation accelerators.<br/></span></li><li><span>How extreme material properties are redefining thermal design requirements —Understand the measurement challenges posed by nanoscale thin films where bulk assumptions fail, engineered ultra-high-conductivity materials (diamond, BAs, BNNTs), and devices operating above 200 °C in wide-band gap systems.</span></li><li><span>Why interfaces and buried layers now govern reliability — Examine how thermal boundary resistance at bonded interfaces, TIM layers, and dielectric stacks has become a first-order reliability accelerator.</span></li><li><span>What a thermal-first design workflow looks like in practice — Learn how measured, scale-appropriate thermal properties can be integrated early in the design cycle to calibrate models, reduce uncertainty, and prevent costly late-stage failures across advanced packaging and 3D architectures.</span></li></ol><div><span><a href="https://content.knowledgehub.wiley.com/heat-beneath-the-surface-thermal-metrology-for-advanced-semiconductor-materials-and-architectures/" target="_blank">Download this free whitepaper now!</a></span></div>]]></description><pubDate>Mon, 23 Mar 2026 10:00:04 +0000</pubDate><guid>https://content.knowledgehub.wiley.com/heat-beneath-the-surface-thermal-metrology-for-advanced-semiconductor-materials-and-architectures/</guid><category>Semiconductors</category><category>Thermal-management</category><category>Scaling</category><category>Type-whitepaper</category><dc:creator>Laser Thermal</dc:creator><media:content medium="image" type="image/png" url="https://assets.rbl.ms/65320713/origin.png"></media:content></item><item><title>What Happens If AI Makes Things Too Easy for Us?</title><link>https://spectrum.ieee.org/frictionless-ai-psychology</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/portrait-of-a-young-white-brunette-woman-behind-her-is-a-collage-of-crumpled-paper-balls-and-ai-sparkle-icons.jpg?id=65324044&width=1200&height=800&coordinates=0%2C208%2C0%2C209"/><br/><br/><p>Most people who regularly use AI tools would say they’re making their lives easier. The technology promises to streamline and take over tasks both professionally and personally—whether that’s summarizing documents, drafting deliverables, generating code, or even offering emotional support. But researchers are concerned AI is making some tasks <em>too</em> easy, and that this will come with unexpected costs.</p><p>In a commentary titled <a href="https://www.nature.com/articles/s44271-026-00402-1" rel="noopener noreferrer" target="_blank"><em>Against Frictionless AI</em></a>, published in <em>Communications Psychology</em><span> on 24 February,</span> psychologists from the University of Toronto discuss what might be lost when AI removes too much effort from human activities. Their argument centers on the idea that friction—difficulty, struggle, and even discomfort—plays an important role in learning, motivation, and meaning. Psychological research has long shown that <a href="https://www.media.mit.edu/publications/your-brain-on-chatgpt/" target="_blank">effortful engagement</a> can deepen understanding and strengthen memory, sometimes described as “desirable difficulties.” <strong></strong></p><p>The authors worry that AI systems capable of instantly producing polished answers or highly responsive conversation may bypass these processes of learning and motivation. By prioritizing outcomes over effort, AI could weaken the experiences that help people develop skills, build relationships, and find meaning in their work.</p><p><em>IEEE Spectrum</em> spoke with the paper’s lead author, <a href="https://www.linkedin.com/in/emily-zohar/?originalSubdomain=ca" target="_blank">Emily Zohar</a>, an experimental psychology Ph.D. student, about why she and her coauthors (psychologists <a href="https://www.psych.utoronto.ca/people/directories/all-faculty/paul-bloom" target="_blank">Paul Bloom</a> and <a href="https://www.utsc.utoronto.ca/psych/person/michael-inzlicht" target="_blank">Michael Inzlicht</a>) argue that friction matters—and what a more human-centered approach to AI design could look like.</p><p><strong>When you say “friction,” what do you mean, from both a cognitive and an interpersonal standpoint?</strong><br/><br/><strong>Zohar:</strong> We define friction as any difficulty encountered during goal pursuit. In the context of work, it involves mental effort—rumination and persistence, staying on a problem for some time, and this helps solidify the idea and the creative process.</p><p>In relationships, friction involves disagreement, compromise, misunderstanding, a back and forth that is natural where you don’t always see eye to eye, and it helps you broaden your horizons. Even the feeling of loneliness is important. It motivates you to find social interactions. So having these negative feelings and difficulty is important in the social context.</p><p><strong>Given that definition, what do you mean by “frictionless” AI?</strong></p><p><strong>Zohar:</strong> Frictionless AI refers to the excessive removal of effort from cognitive and social tasks. With AI, as we typically use it, it’s really easy to go from ideation right to the end product. You ask AI to solve something with one prompt, and it completes the whole thing. This is a problem because it takes away the intermediate steps that really drive motivation and learning, and it prioritizes outcome over process. Rather than working through the steps, AI does that meaningful work for you.<br/><br/>There’s a lot of research showing <a href="https://arxiv.org/abs/2409.14511" target="_blank">work products</a> are better with AI. That makes sense, it has all this knowledge, but it does worry us as it may be eroding something essential that will have long-term consequences. If you’re faced with the same problem and AI is removed, you don’t have the required knowledge to know how to face the problem next time.</p><p><strong>You argue that removing friction can harm learning and relationships. What role do effort and struggle play in human development?</strong></p><p><strong>Zohar: </strong>In learning, the term is “desirable difficulties.” It’s the idea of effort and work, not just any effort but <em>manageable</em> effort. Facing problems that you can overcome, but you have to work at them a bit, that’s the key idea of friction. We don’t want you to face insurmountable problems. We want you to work hard, but still be able to overcome it. This helps you really digest information and learn from it.</p><p>In interpersonal relationships, you have to face some difficulties to see other perspectives and learn from them, and learn to be accepting of others. If you’re used to an AI reinforcing all your ideas and being sycophantic, you’ll come into the real world and you won’t be used to seeing other ideas. You won’t know how to interact socially because you’ll expect people to always be on your side and agree with you. You won’t learn that life doesn’t always go exactly how you expect it to, and conversations don’t always go the way you want them to.</p><h2>AI’s Impact on Creative Processes</h2><p><strong>A lot of technologies have historically aimed to reduce effort: calculators, washing machines, spell-check. What’s different about AI?</strong></p><p><strong>Zohar:</strong> Past technologies have mostly focused on reducing physical effort. We don’t have to go down to the lake to wash our laundry anymore. [Past technologies] took away the mundane tasks that weren’t driving our learning and growth, they were just adding unneeded obstacles and taking away time from more important tasks.</p><p>But AI is taking away effort from creative and cognitive processes that drive meaning, motivation, and learning. That’s a key difference, because it’s not taking away friction from tasks that don’t serve us. It’s taking away friction from experiences that are really important and integral to our development.</p><p><strong>Are there contexts where AI is already removing beneficial friction? How might the impacts of reduced friction show up over time?</strong></p><p><strong>Zohar:</strong> One clear example is writing. People increasingly rely on AI to draft everything from emails to essays, removing many instances of beneficial friction. Research shows that people trust responses less when they learn they were written by AI, judge AI-generated products as less creative and less valuable, and have greater difficulty remembering their own work products when they were produced with AI assistance. Outsourcing writing to AI strips away both social and cognitive friction.</p><p><a data-linked-post="2671645555" href="https://spectrum.ieee.org/vibe-coding" target="_blank">Vibe coding</a> is another good example. If you’re a programmer, coding is integral to what drives your meaning. People get meaning out of their work, and if you’re substituting that with AI, it could be detrimental. The negative impact of frictionless AI is that it takes away friction from things that are really important to who you are as a person, and your skills.</p><p>One area I worry about a lot is <a data-linked-post="2656019975" href="https://spectrum.ieee.org/kids-ai" target="_blank">adolescents using AI in general</a>. It’s a really important developmental period to learn and grow and find the path you’ll follow. So if you don’t have these effortful interactions with work and relationships that teach you how to think, this will have long-term detrimental impacts. They might not be able to think critically in the same way, because they never had to before. If they’re turning to AI for social relationships at such a young age, that could really erode important skills they should be learning at that age.</p><p><strong>What is productive friction?</strong></p><p><strong>Zohar:</strong> Friction goes along a continuum. With too little friction, you’re not getting learning and motivation. Too much friction and the task becomes overwhelming. Productive friction falls right in the middle, where struggle leads to achievement. It’s effortful but possible, and it requires you to think critically and work on a problem for some time or face some difficulty in the process.</p><p>An example we used in the paper is the difference between taking a chairlift and hiking up a mountain. They both get to the top, but with the chairlift, you don’t get any growth benefits, while the hiker’s climb involves difficulties and a sense of achievement. It becomes much more of an experience and a learning opportunity versus the person who just went up the chairlift effortlessly.</p><p><strong>Do you envision AI that sometimes deliberately slows people down or asks them to do part of the work themselves?</strong></p><p><strong>Zohar:</strong> It’s important in behavioral science to think about the default option, because people don’t usually change their default. So right now, the default in AI is to give you your answer and probe you to keep going down the rabbit hole. But I think we could think about AI in a different way. Maybe we can make the default more constructive. Instead of just jumping to the answer, it’s more of a process model where it helps you think about the problem and teaches you along the way, so it’s more collaborative rather than a one-stop shop for the answer.</p><p><strong>How might users of these systems and the companies developing them feel about such a design shift?</strong></p><p><strong>Zohar: </strong>For the makers of these systems, the biggest concern is the pushback. People are used to going in and just getting the answer, and they might be really resistant to a design that makes them work more for it. But it might feed more engagement, because you have to go back and forth and find the answer together.</p><p>Ultimately I think it has to come from the companies making these models, if they think [a more friction-full design] would help people. Friction-full AI is more of a long-term product. It’s hard to say if that would motivate companies to change their models to include moderate friction. But in the long term, I think this would be beneficial.</p>]]></description><pubDate>Sun, 22 Mar 2026 13:00:04 +0000</pubDate><guid>https://spectrum.ieee.org/frictionless-ai-psychology</guid><category>Psychology</category><category>Artificial-intelligence</category><category>Cognitive-science</category><dc:creator>Vanessa Bates Ramirez</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/portrait-of-a-young-white-brunette-woman-behind-her-is-a-collage-of-crumpled-paper-balls-and-ai-sparkle-icons.jpg?id=65324044&amp;width=980"></media:content></item><item><title>AI Aims for Autonomous Wheelchair Navigation</title><link>https://spectrum.ieee.org/autonomous-smart-wheelchair</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/side-view-of-a-power-wheelchair-equipped-with-a-padded-bucket-seat-and-tablet-sized-monitor-below-a-computer-generated-maps-of.jpg?id=65316423&width=1200&height=800&coordinates=0%2C208%2C0%2C209"/><br/><br/><p>Wheelchair users with severe disabilities can often navigate tight spaces better than most robotic systems can. <span>A wave of new smart-wheelchair research, including findings presented in Anaheim, Calif., earlier this month, is now testing whether AI-powered systems can, or should, fully close this gap.</span></p><p><a href="https://user.informatik.uni-bremen.de/cmandel/" target="_blank">Christian Mandel</a>—senior researcher at the <a href="https://www.dfki.de/en/web" target="_blank">German Research Center for Artificial Intelligence</a> (DFKI) in Bremen, Germany—<span>co-led a research team together with his colleague <a href="https://user.informatik.uni-bremen.de/autexier/index.php" target="_blank">Serge Autexier</a></span><span> that developed prototype sensor-equipped electric wheelchairs designed to navigate a roomful of potential obstacles. The researchers also tested a new safety system that integrated sensor data from the wheelchair and from sensors in the room, including from </span><a href="https://spectrum.ieee.org/tag/drones" target="_self">drone</a><span>-based </span>color and depth cameras<span>.</span></p><p>Mandel says the team’s smart wheelchairs were both semiautonomous and autonomous.</p><p>“Semiautonomous is the shared control system where the person sitting in the wheelchair uses the joystick to drive,” Mandel says. “Fully autonomous is controlled by natural-language input. You say, ‘Please drive me to the coffee machine.’ ”<a href="#_msocom_2" target="_blank"></a></p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="Close-up of a thin rectangular camera installed underneath an electric wheelchair's joystick controller." class="rm-shortcode" data-rm-shortcode-id="dbd1d07a3ba04703e3cd78d8f2980624" data-rm-shortcode-name="rebelmouse-image" id="d4669" loading="lazy" src="https://spectrum.ieee.org/media-library/close-up-of-a-thin-rectangular-camera-installed-underneath-an-electric-wheelchair-s-joystick-controller.jpg?id=65317537&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">This is a close-up of the wheelchair’s joystick and camera.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">DFKI</small></p><p>The researchers conducted experiments (<a target="_blank">part of a larger project called the </a><a href="https://www.dfki.de/en/web/research/projects-and-publications/project/rexasi-pro" target="_blank">Reliable and Explainable Swarm Intelligence for People With Reduced Mobility</a>, or REXASI-PRO) using two identical smart wheelchairs that each contained two lidars, a 3D camera, odometers, user interfaces, and an embedded computer.</p><p>In contrast to semiautonomous mode, where the participant controls the wheelchair with a joystick, in autonomous mode, control involves the open-source <a href="https://roboticsbackend.com/ros2-nav2-tutorial/" target="_blank">ROS2 Nav2</a> navigation system using natural-language input.  The wheelchairs also used simultaneous localization and mapping (<a href="https://en.wikipedia.org/wiki/Simultaneous_localization_and_mapping" rel="noopener noreferrer" target="_blank">SLAM</a>) maps and local obstacle-avoidance motion controllers.</p><p>One scenario that Mandel and his team tested involved the user pressing a key on the wheelchair’s human-machine interface, speaking a command, then confirming or rejecting the instruction via that same interface. Once the user confirmed the command, the mobility device guided the user along a path to the destination, while sensors attempted to detect obstacles in the way and adjust the mobility device accordingly to avoid them.</p><h3>When Are Smart Wheelchairs Bad Value?</h3><p>According to Pooja Viswanathan, CEO & founder of the <a rel="noopener noreferrer" target="_blank">Toronto-based</a> Braze Mobility, research in the field of mobile assistive technology should also prioritize keeping these devices readily available to everyday consumers.</p><p>“Cost remains a major barrier,” she says. “Funding systems are often not designed to support advanced add-on intelligence unless there is very clear evidence of value and safety. Reliability is another barrier. A smart wheelchair has to work not just in ideal conditions, but in the messy, variable conditions of daily life. And there is also the human factors dimension. Users have different cognitive, motor, sensory, and environmental needs, so one solution rarely fits all.”</p><p>For its part, Braze makes <a href="https://brazemobility.com/" rel="noopener noreferrer" target="_blank">blind-spot sensors</a> for electric wheelchairs. The sensors detect obstacles in areas that can be difficult for a user to see. The sensors can also be added to any wheelchair to transform it into a smart wheelchair by providing multimodal alerts to the user. This approach attempts to support users rather than replace them.</p><p>According to Louise Devinge, a biomedical research engineer from <a href="https://en.wikipedia.org/wiki/Research_Institute_of_Computer_Science_and_Random_Systems" rel="noopener noreferrer" target="_blank">IRISA</a> (Research Institute of Computer Science and Random Systems) in Rennes, France, the increased complexity of smart wheelchairs demands more sensing. And that requires careful management of communication and synchronization within the wheelchair’s system. “The more sensing, computation, and autonomy you add,” she says, “the harder it becomes to ensure robust performance across the full range of real-world environments that wheelchair users encounter.”</p><p>In the near term, in other words, the field’s biggest challenge is not about replacing the wheelchair user with AI smarts but rather about designing better partnerships between the user and the technology.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Rendering of an electric wheelchair moving towards a wall. The chair is divided into four ground-parallel quadrants that each represent a different safety zone where intersections with obstacles are checked. At the same height as these quadrants, are four lines on the wall that represent virtual laser scans.  " class="rm-shortcode" data-rm-shortcode-id="1834a2b260ce3cf9c3bd1863293c4d99" data-rm-shortcode-name="rebelmouse-image" id="f0d73" loading="lazy" src="https://spectrum.ieee.org/media-library/rendering-of-an-electric-wheelchair-moving-towards-a-wall-the-chair-is-divided-into-four-ground-parallel-quadrants-that-each-re.jpg?id=65316452&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">This image shows data representations used by the 3D Driving Assistant. These include immutable sensor percepts such as laser scans and point clouds, as well as derived representations like the virtual laser scans and grid maps. Finally, the robot shape collection describes the wheelchair’s physical borders at different heights.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">DFKI</small></p><h3>Where Will Smart Wheelchairs Go From Here?</h3><p>Mandel says he expects to see smart wheelchairs ready for the mainstream marketplace within 10 years.</p><p>Viswanathan says the REXASI-PRO system, while out of reach of <a href="https://spectrum.ieee.org/assistive-technology-lidar-wheelchair" target="_self">present-day smart wheelchair technologies</a>, is important for the longer term. “It reflects the more ambitious end of the smart wheelchair spectrum,” she says. “Its strengths appear to lie in intelligent navigation, advanced sensing, and the broader effort to build a wheelchair that can interpret and respond to complex environments in a more autonomous way. From a research standpoint, that is exactly the kind of work that pushes the field forward. It also appears to take seriously the importance of trustworthy and explainable AI, which is essential in any mobility technology where safety, reliability, and user confidence are paramount.”</p><p>Mandel says he’s ultimately in pursuit of the inspiration that got him into this field years ago. As a young researcher, he says, he helped develop a smart wheelchair system controllable with a head joystick.</p><p>However, Mandel says he realized after many trials that the smart wheelchair system he was working on had a long way to go because, as he says, “at that point in time, I realized that even persons that had severe handicaps [traveling through] a narrow passage, they did very, very well.</p><p>“And then I realized, okay, there is this need for this technology, but never underestimate what [wheelchair users] can do without it.”</p><p><a target="_blank">The DFKI researchers presented </a><a href="https://www.dfki.de/en/web/research/projects-and-publications/publication/16538" target="_blank">their work</a> earlier this month at the <a href="https://conference.csun.at/event/2026/session-schedule" target="_blank">CSUN Assistive Technology Conference</a> in Anaheim, Calif.</p><p><em>This article was supported by the <a href="https://spectrum.ieee.org/tag/ieee-foundation" target="_self">IEEE Foundation</a> and a Jon C. Taenzer fellowship grant.</em></p>]]></description><pubDate>Fri, 20 Mar 2026 18:49:12 +0000</pubDate><guid>https://spectrum.ieee.org/autonomous-smart-wheelchair</guid><category>Wheelchairs</category><category>Taenzer-fellowship</category><category>Navigation</category><category>Artificial-intelligence</category><dc:creator>Jason Hahr</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/side-view-of-a-power-wheelchair-equipped-with-a-padded-bucket-seat-and-tablet-sized-monitor-below-a-computer-generated-maps-of.jpg?id=65316423&amp;width=980"></media:content></item><item><title>Startups Bring Optical Metamaterials to AI Data Centers</title><link>https://spectrum.ieee.org/optical-metamaterials-ai-data-centers</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-hand-holding-a-microchip-between-thumb-and-forefinger.jpg?id=65322426&width=1200&height=800&coordinates=0%2C83%2C0%2C84"/><br/><br/><p><span>Light-warping physics made “invisibility cloaks” a possibility. Now two startups hope to harness the science underlying this advance to boost the bandwidth of data centers and speed artificial intelligence.</span></p><p>Roughly 20 years ago, scientists developed the <a href="https://www.science.org/doi/10.1126/science.1125907" target="_blank">first</a> <a href="https://www.science.org/doi/10.1126/science.1133628" target="_blank"> structures</a> capable of curving light around objects to conceal them. These are composed of optical <a href="https://spectrum.ieee.org/two-photon-lithography-3d-printing" target="_self">metamaterials</a>—materials with structures smaller than the wavelengths they are designed to manipulate, letting them bend light in unexpected ways.</p><p>The problem with optical cloaks? “There’s no market for them,” says Patrick Bowen, cofounder and CEO of photonic computing startup <a href="https://www.neurophos.com/" target="_blank">Neurophos</a> in Austin, Texas. For instance, each optical cloak typically works only on a single color of light instead of on all visible colors as you might want for stealth applications.</p><p>Now companies are devising more practical uses for the science behind cloaks, such as improving the switches that connect computers in data centers for AI and other cloud services. Increasingly, <a href="https://newsletter.semianalysis.com/p/google-apollo-the-3-billion-game" target="_blank">data centers are looking to use optical circuit switches </a>to overcome the bandwidth limits and power consumption of conventional electronic switches and networks that require converting data between light to electrons multiple times.</p><p class="ieee-inbody-related">RELATED:  <a href="https://spectrum.ieee.org/optical-interconnects-imec-silicon-photonics" target="_blank">Semiconductor Industry Closes in on 400 Gb/s Photonics Milestone</a></p><p>However, today’s optical switching technologies have drawbacks of their own. For instance, ones that depend on silicon photonics face problems with energy efficiency, while those that rely on <a href="https://spectrum.ieee.org/self-assembly" target="_self">microelectromechanical systems (MEMS)</a> can prove unreliable, says Sam Heidari, CEO of optical metasurface startup <a href="https://lumotive.com/" rel="noopener noreferrer" target="_blank">Lumotive</a> in Redmond, Wash.</p><p>Instead, <a href="https://www.nature.com/articles/s44287-024-00136-4" rel="noopener noreferrer" target="_blank">Lumotive has developed metamaterials with adjustable properties</a>. Its new microchip, which debuted 19 March, is covered with copper structures built using standard chipmaking techniques. Between these copper features are <a href="https://spectrum.ieee.org/metasurface-displays" target="_self">liquid crystal</a> elements. The structure of these elements are electronically programmable, just like in liquid crystal displays (LCDs), to alter the optical properties of the metamaterial chip.</p><p>The microchip can precisely steer, lens, shape, and split beams of light reflected off its surface. It can perform all the same functions as multiple optical components with no moving parts in a programmable way in real time, according to Lumotive. “Having no moving parts significantly improves reliability,” Heidari says.</p><p>“We had to go through a lot of R&D at the foundries to not only make our devices functional, but also commercially viable in terms of the right cost and right reliability,” Heidari says.</p><p>The company says its new chips are capable of manipulating not only the industry’s standard of 256 by 256 ports, but could scale up to 10,000 by 10,000. “We think this is game-changing for data centers,” Heidari says. Lumotive plans to launch its first optical switches at the end of 2026.</p><h2>Optical Computing With Metamaterials</h2><p>Similarly, Neurophos hopes its technology may be transformative for artificial intelligence. Since AI is proving energy hungry when run on conventional electronics, scientists are exploring <a href="https://spectrum.ieee.org/optical-neural-networks" target="_self">optical computing</a> as a low-power alternative by processing data with light instead of electrons.</p><p>However, optical processors in the works today are typically far too bulky to achieve a compute density competitive with the best modern electronic processors, Bowen says. Neurophos says it can use metamaterials to build optical modulators—the optical equivalent of a transistor—that are 1/10,000th the size of today’s designs using standard chipmaking processes. “It’s entirely CMOS,” Bowen says. “There are no exotic materials in it.”</p><p>When a laser beam encoding data shines on a Neurophos chip, the way in which each metamaterial element is configured alters the reflected beam to encode results from complex AI tasks. “We basically fit a 1,000- by-1,000 array of optical modulators on a tiny 5-by-5-millimeter area on a chip,” Bowen says. “If you wanted to do that with off-the-shelf silicon photonics, your chip would be a square meter in size.”</p><p>All in all, Bowen claims the Neurophos microchip will offer 50 times greater compute density and 50 times greater energy efficiency than Nvidia’s Blackwell-generation GPU. The company says that hyperscalers—the world’s biggest cloud service providers—will evaluate two upcoming proof-of-concept chips this year. Neurophos is targeting its first systems for early 2028, with production ramping mid-2028.</p>]]></description><pubDate>Thu, 19 Mar 2026 19:19:43 +0000</pubDate><guid>https://spectrum.ieee.org/optical-metamaterials-ai-data-centers</guid><category>Artificial-intelligence</category><category>Data-center</category><category>Optical-switch</category><category>Optical-computing</category><category>Metamaterial</category><category>Metamaterials</category><dc:creator>Charles Q. Choi</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-hand-holding-a-microchip-between-thumb-and-forefinger.jpg?id=65322426&amp;width=980"></media:content></item><item><title>How Your Virtual Twin Could One Day Save Your Life</title><link>https://spectrum.ieee.org/living-heart-project-virtual-twins</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/two-color-coded-computer-simulations-of-a-human-heart-the-simulation-on-the-left-shows-the-muscle-structure-and-the-simulation.png?id=65278129&width=1200&height=800&coordinates=0%2C82%2C0%2C82"/><br/><br/><p><strong>One morning in May </strong>2019, a cardiac surgeon stepped into the operating room at Boston Children’s Hospital more prepared than ever before to perform a high-risk procedure to rebuild a child’s heart. The surgeon was experienced, but he had an additional advantage: He had already performed the procedure on this child dozens of times—virtually. He knew exactly what to do before the first cut was made. Even more important, he knew which strategies would provide the best possible outcome for the child whose life was in his hands.</p><p>How was this possible? Over the prior weeks, the hospital’s surgical and cardio-engineering teams had come together to build a fully functioning model of the child’s heart and surrounding vascular system from MRI and CT scans. They began by carefully converting the medical imaging into a 3D model, then used physics to bring the 3D heart to life, creating a dynamic <a href="https://spectrum.ieee.org/virtual-hearts-improve-cardiac-surgery" target="_self">digital replica</a> of the patient’s physiology. The mock-up reproduced this particular heart’s unique behavior, including details of blood flow, pressure differentials, and muscle-tissue stresses.</p><p>This type of model, known as a virtual twin, can do more than identify medical problems—it can provide detailed diagnostic insights. In Boston, the team used the model to predict how the child’s heart would respond to any cut or stitch, allowing the surgeon to test many strategies to find the best one for this patient’s exact anatomy.</p><p>That day, the stakes were high. With the patient’s unique condition—a heart defect in which large holes between the atria and ventricles were causing blood to flow between all four chambers—there was no manual or textbook to fully guide the doctors. The condition strains the lungs, so the doctors planned an open-heart surgery to reroute deoxygenated blood from the lower body directly to the lungs, bypassing the heart. Typically with this kind of surgery, decisions would be made on the fly, under demanding conditions, and with high uncertainty. But in this case, the plan had been tested in advance, and the entire team had rehearsed it before the first incision. The surgery was a complete success.</p><p>Such procedures have become routine at the Boston hospital. Since that first patient, nearly 2,000 procedures have been guided by virtual-twin modeling. This is the power of the technology behind the <a href="https://www.3ds.com/3dexperiencelab/portfolio/living-heart" rel="noopener noreferrer" target="_blank">Living Heart Project</a>, which I launched in 2014, five years before that first procedure. The project started as an exploratory initiative to see if modeling the human heart was possible. Now with more than 150 member organizations across 28 countries, the project includes dozens of multidisciplinary teams that regularly use multiscale virtual twins of the heart and other vital organs.</p><p>This technology is reshaping how we understand and treat the human body. To reach this transformative moment, we had to solve a fundamental challenge: building a digital heart accurate enough—and trustworthy enough—to guide real clinical decisions.</p><h2>A father’s concern</h2><p>Now entering its second decade, the Living Heart Project was born in part from a personal conviction. For many years, I had watched helplessly as my daughter Jesse faced endless diagnostic uncertainty due to a <a href="https://doi.org/10.1016/B978-1-4557-0599-3.00039-9" rel="noopener noreferrer" target="_blank">rare congenital heart condition</a> in which the position of the ventricles is reversed, threatening her life as she grew. As an engineer, I understood that the heart was an array of pumping chambers, controlled by an electrical signal and its blood flow carefully regulated by valves. Yet I struggled to grasp the unique structure and behavior of my daughter’s heart well enough to contribute meaningfully to her care. Her specialists knew the bleak forecast children like her faced if left untreated, but because every heart with her condition is anatomically unique, they had little more than their best guesses to guide their decisions about what to do and when to do it. With each specialist, a new guess.</p><p>Then my engineering curiosity sparked a question that has guided my career ever since: Why can’t we simulate the human body the way we <a href="https://spectrum.ieee.org/selfdriving-cars-learn-about-road-hazards-through-augmented-reality" target="_self">simulate a car</a> or a plane?</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="woman facing away and looking at a wall where the simulated interior of a heart is projected" class="rm-shortcode" data-rm-shortcode-id="442abe00bb6d81b4be0ad13e4ec3880e" data-rm-shortcode-name="rebelmouse-image" id="09f25" loading="lazy" src="https://spectrum.ieee.org/media-library/woman-facing-away-and-looking-at-a-wall-where-the-simulated-interior-of-a-heart-is-projected.png?id=65301974&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">At a visualization center in Boston, VR imagery helps the mother of a young girl with a complex heart defect understand the inner workings of her child’s heart. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Dassault Systèmes</small></p><p>I had spent my career developing powerful computational tools to help engineers build digital models of complex mechanical systems, using models that ranged from the interactions of individual atoms to the components of entire vehicles. What most of these models had in common was the use of physics to predict behavior and optimize performance. But in medicine today, those same physics-based approaches rarely inform decision-making. In most clinical settings, treatment decisions still hinge on judgments drawn from static 2D images, statistical guidelines, and retrospective studies.</p><p>This was not always the case. Historically, physics was central to medicine. The word “physician” itself traces back to the Latin <em><em>physica</em></em>, which translates to “natural science.” Early doctors were, in a sense, applied physicists. They understood the heart as a pump, the lungs as bellows, and the body as a dynamic system. To be a physician meant you were a master of physics as it applied to the human body.</p><p>As medicine matured, biology and chemistry grew to dominate the field, and the knowledge of physics got left behind. But for patients like my daughter, that child in Boston, and millions like them, outcomes are governed by mechanics. No pill or ointment—no chemistry-based solution—would help, only physics. While I did not realize it at the time, virtual twins can reunite modern physicians with their roots, using engineering principles, simulation science, and artificial intelligence.</p><h2>A decade of progress</h2><p>The LHP concept was simple: Could we combine what hundreds of experts across many specialties knew about the human heart to build a digital twin accurate enough to be trusted, flexible enough to personalize, and predictive enough to guide clinical care?</p><p>We invited researchers, clinicians, device and drug companies, and government regulators to share their data, tools, and knowledge toward a common goal that would lift the entire field of medicine. The Living Heart Project launched with a dozen or so institutions on board. Within a year, we had created the first fully functional virtual twin of the human heart.</p><p>The Living Heart was not an anatomical rendering, tuned to simply replicate what we observed. It was a first-principles model, coupling the network of fibers in the <a href="https://spectrum.ieee.org/medtronics-cardioinsight-electrode-vest-maps-hearts-electrical-system" target="_self">heart’s electrical system</a>, the biological battery that keeps us alive, with the heart’s mechanical response, the muscle contractions that we know as the heartbeat.</p><p class="shortcode-media shortcode-media-youtube"> <span class="rm-shortcode" data-rm-shortcode-id="85d721660928d134fc0039fb17d76716" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/ae_IqlxgCME?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span> <small class="image-media media-caption" placeholder="Add Photo Caption...">The Living Heart virtual twin simulates how the heart beats, offering different views to help scientists and doctors better predict how it will respond to disease or treatment. The center view shows the fine engineering mesh, the detailed framework that allows computers to model the heart’s motion. The image on the right uses colors to show the electrical wave that drives the heartbeat as it conducts through the muscle, and the image on the left shows how much strain is on the tissue as it stretches and squeezes. </small> <small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Dassault Systèmes</small> </p><p>Academic researchers had long explored computational models of the heart, but those projects were typically limited by the technology they had access to. Our version was built on industrial-grade simulation software from <a href="https://www.3ds.com/" target="_blank">Dassault Systèmes</a>, a company best known for modeling tools used in aerospace and automotive engineering, where I was working to develop the engineering simulation division. This platform gave teams the tools to personalize an individual heart model using the patient’s MRI and CT data, blood-pressure readings, and echocardiogram measurements, directly linking scans to simulations.</p><p>Surgeons then began using the Living Heart to model procedures. Device makers used it to design and test implants. Pharmaceutical companies used it to evaluate drug effects such as toxicity. Hundreds of publications have emerged from the project, and because they all share the same foundation, the findings can be reproduced, reused, and built upon. With each application, the research community’s understanding of the heart snowballed.</p><p>Early on, we also addressed an essential requirement for these innovations to make it to patients: regulatory acceptance. Within the project’s first year, the U.S Food and Drug Administration <a href="https://www.3ds.com/newsroom/press-releases/dassault-systemes-signs-research-agreement-food-and-drug-administration-its-living-heart-project" target="_blank">agreed to join the project</a> as an observer. Over the next several years, methods for using virtual-heart models as scientific evidence began to take shape within regulatory research programs. In 2019, we formalized a second five-year collaboration with the FDA’s Center for Devices and Radiological Health with a specific goal.</p><p>That goal was to use the heart model to create a virtual patient population and re-create a pivotal trial of a previously approved device for repairing the heart’s mitral valve. This helped our team learn how to create such a population, and let the FDA experiment with evaluating virtual evidence as a replacement for evidence from flesh-and-blood patients. In August 2024, we <a href="https://pubmed.ncbi.nlm.nih.gov/39188879/" target="_blank">published the results</a>, creating the first FDA-led guidelines for in silico clinical trials and establishing a new paradigm for streamlining and reducing risk in the entire clinical-trial process.</p><p>In 10 years, we went from a concept that many people doubted could be achieved to regulatory reality. But building the heart was only the beginning. Following the template set by the heart team, we’ve expanded the project to develop virtual twins of other organs, including the lungs, liver, brain, eyes, and gut. Each corresponds to a different medical domain, which has its own community, data types, and clinical use cases. Working independently, these teams are progressing toward a breakthrough in our understanding of the human body: a multiscale, modular twin platform where each organ twin could plug into a unified virtual human.</p><h2>How a digital twin of the heart is constructed</h2><p>A cardiac digital twin starts with medical imaging, typically MRI, CT, or both. The slices are reconstructed into the 3D geometry of the heart and connected vessels. The geometry of the whole organ must then be segmented into its constituent parts, so each substructure—atria, ventricles, valves, and so on—can be assigned their unique properties.</p><p>At this point, the object is converted to a functional, computational model that can represent how the various cardiac tissues deform under load—the mechanics. The complete digital twin model becomes “living” when we integrate the electrical fiber network that drives mechanical contractions in the muscle tissue.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="two computer simulations of a heart. The simulation on left shows the left ventricle with a triangular grid across the 3D surface. The simulation on right shows the exterior of a heart including vasculature and fat. " class="rm-shortcode" data-rm-shortcode-id="8b175dd3f95e87ac7f36ab39b38f9784" data-rm-shortcode-name="rebelmouse-image" id="deda7" loading="lazy" src="https://spectrum.ieee.org/media-library/two-computer-simulations-of-a-heart-the-simulation-on-left-shows-the-left-ventricle-with-a-triangular-grid-across-the-3d-surfac.png?id=65301904&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Each part of the heart, such as the left ventricle [left], is superimposed with a detailed digital mesh to re-create its physiology. These pieces come together to form an anatomically accurate rendering of the whole organ [right].</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Dassault Systèmes</small></p><p>To simulate circulation, the twin adds computational models of hemodynamics, the physics of blood flow and pressure. The model is constrained by boundary conditions of blood flow, valve behavior, and vascular resistance set to closely match human physiology. This lets the model predict blood flow patterns, pressure differentials, and tissue stresses.</p><p>Finally, the model is personalized and calibrated using available patient data, such as how much the volume of the heart chambers changes during the cardiac cycle, pressure measurements, and the timing of electrical pulses. This means the twin reflects not only the patient’s anatomy but how their specific heart functions.</p><h2>Building bigger cohorts with generative AI</h2><p>When the <a href="https://discover.3ds.com/fda-enrichment-clinical-trial" target="_blank">FDA in silico clinical trial initiative</a> launched in 2019, the project’s focus shifted from these handcrafted virtual twins of specific patients to cohorts large enough to stand in for entire trial populations. That scale is feasible today only because virtual twins have converged with generative AI. Modeling thousands of patients’ responses to a treatment or projecting years of disease progression is prohibitively slow with conventional digital-twin simulations. Generative AI removes that bottleneck.</p><p>AI boosts the capability of virtual twins in two complementary ways. First, machine learning algorithms are unrivaled at integrating the patchwork of imaging, sensor, and clinical records needed to build a high-fidelity twin. The algorithms rapidly search thousands of model permutations, benchmark each against patient data, and converge on the most accurate representation. Workflows that once required months of manual tuning can now be completed in days, making it realistic to spin up population-scale cohorts or to personalize a single twin on the fly in the clinic.</p><p>Second, enriching AI models’ training sets with data from validated virtual patients grounds the AI simulations in physics. By contrast, many conventional AI predictions for patient trajectories rely on statistical modeling trained on retrospective datasets. Such models can drift beyond physiological reality, but virtual twins anchor predictions in the laws of hemodynamics, electrophysiology, and tissue mechanics. This added rigor is indispensable for both research and clinical care—especially in areas where real-world data are scarce, whether because a disease is rare or because certain patient populations, such as children, are underrepresented in existing datasets.</p><h2>Enabling in silico clinical trials</h2><p>On the research side, the FDA-sponsored In Silico Clinical Trial Project that we completed in 2024 opened a new world for medical innovations. A conventional clinical trial may take a decade, and 90 percent of new drug treatments fail in the process. Virtual twins, combined with AI methods, allow researchers to design and test treatments quickly in a simulated human environment. With a small library of virtual twins, AI models can rapidly create expansive virtual patient cohorts to cover any subset of the general population. As clinical data becomes available, it can be added into the training set to increase reliability and enable better predictions.</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="3D simulations of the brain, foot, and lungs. A quadrant of the brain is cut out, showing a dense network of connections between color-coded sections of the brain. The foot shows a gray outline of bones and points of soft tissue strain in red at the ankle and heel. In the lung model, the trachea is colored green flowing into blue bronchi. " class="rm-shortcode" data-rm-shortcode-id="6c65f028c501081d47120dbb37f2d816" data-rm-shortcode-name="rebelmouse-image" id="90af6" loading="lazy" src="https://spectrum.ieee.org/media-library/3d-simulations-of-the-brain-foot-and-lungs-a-quadrant-of-the-brain-is-cut-out-showing-a-dense-network-of-connections-between.png?id=65302220&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">The Living Heart Project has expanded beyond the heart, modeling organs throughout the body. The 3D brain reconstruction [top] shows major pathways in the brain’s white matter connecting color-coded regions of the brain. The lung virtual twin [middle] combines the organ’s geometry with a physics-based simulation of air flowing down the trachea and into the bronchi. And the cross section of a patient’s foot [bottom] shows points of strain in the soft tissue when bearing weight. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Dassault Systèmes</small></p><p>Virtual twin cohorts can represent a realistic population by building individual “virtual patients” that vary by age, gender, race, weight, disease state, comorbidities, and lifestyle factors. These twins can be used as a rich training set for the AI model, which can expand the cohort from dozens to hundreds of thousands. Next the virtual cohort can be filtered to identify patients likely to respond to a treatment, increasing the chances of a successful trial for the target population.</p><p>The trial design can also include a sampling of patient types less likely to respond or with elevated risk factors, thus allowing regulators and clinicians to understand the risks to the broader population without jeopardizing overall trial success. This methodology enhances precision and efficiency in clinical research, providing population-level insights previously available only after many years of real-world evidence.</p><p>Of course, though today’s heart digital twins are powerful, they’re not perfect replicas. Their accuracy is bounded by three main factors: what we can measure (for example, image resolution or the uncertainty of how tissue behaves in real life), what we must assume about the physiology, and what we can validate against real outcomes. Many inputs, like scarring, microvascular function, or drug effects are difficult to capture clinically, so models often rely on population data or indirect estimation. That means predictions can be highly reliable for certain questions but remain less certain for others. Additionally, today’s digital twins lack validation for predicting long-term outcomes years in the future, because the technology has been in use for only a few years.</p><p>Over time, each of these limitations will steadily shrink. Richer, more standardized data will tighten personalization of the models. AI tools will help automate labor-intensive steps. And the collection of longitudinal data will improve the model’s ability to reliably predict how the body will evolve over time.</p><h2>How virtual twins will change health care</h2><p>Throughout modern medicine, new technologies have sharpened our ability to <a href="https://spectrum.ieee.org/ai-doctor" target="_self">diagnose</a>, providing ever-clearer images, lab data, and analytics that tell physicians what is presently happening inside a patient’s body. Virtual twins shift that paradigm, giving clinicians a predictive tool.</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" rel="float: left;" style="float: left;"> <img alt="gif of a lung simulation. The lungs are blue when deflated then grow and become green with points of red. " class="rm-shortcode" data-rm-shortcode-id="99cdfc0b66a34d7bf081125259464d73" data-rm-shortcode-name="rebelmouse-image" id="499fe" loading="lazy" src="https://spectrum.ieee.org/media-library/gif-of-a-lung-simulation-the-lungs-are-blue-when-deflated-then-grow-and-become-green-with-points-of-red.gif?id=65302107&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">This “Living Lung” virtual-twin simulation shows strain patterns during breathing. </small> <small class="image-media media-photo-credit" placeholder="Add Photo Credit...">            Mona Eskandari/UC Riverside        </small> </p><p>Early demonstrations are already appearing in many areas of medicine, including cardiology, orthopedics, and oncology. Soon, doctors will also be able to collaborate across specialties, using a patient-specific virtual twin as the common ground for discussing potential interactions or side effects they couldn’t predict independently.</p><p>Although these applications will take some time to become the standard in clinical care, more changes are on the horizon. Real-time <a href="https://spectrum.ieee.org/wearable-health-data-standards" target="_self">data from wearables</a>, for example, could continuously update a patient’s personalized virtual twin. This approach could empower patients to understand and engage more deeply in their care, as they could see the direct effects of medical and lifestyle changes. In parallel, their doctors could get comprehensive data feeds, using virtual twins to monitor progress.</p><p><span>Imagine a digital companion that shows how your particular heart will react to different amounts of salt intake, stress, or sleep deprivation. Or a visual explanation of how your upcoming surgery will affect your circulation or breathing. Virtual twins could demystify the body for patients, fostering trust and encouraging proactive health decisions.</span></p><h3>How are virtual twins being used in medicine?</h3><br/><ul><li>Virtual twins have guided <strong>cardiovascular surgeries</strong>, providing predictions and exposing hidden details that even expert clinicians might miss, such as subtle tissue responses and flow dynamics.<br/></li><li><strong>Oncologists</strong> are modeling tumor growth and the body’s response to different therapies, reducing the uncertainty in choosing the best treatment path for both medical and quality-of-life metrics.<br/></li><li><strong>Orthopedic</strong> specialists are personalizing implants to deliver custom-made solutions, considering not only the local environment but also the overall body kinematics that will govern long-term outcomes.</li></ul><h2>A new era of healing</h2><p>With the Living Heart Project, we’re bringing physics back to physicians. Modern physicians won’t need to be physicists, any more than they need to be chemists to use pharmacology. However, to benefit from the new technology, they will need to adapt their approach to care.</p><p>This means no longer seeing the body as a collection of discrete organs and considering only symptoms, but instead viewing it as a dynamic system that can be understood, and in most cases, guided toward health. It means no longer guessing what might work but knowing—because the simulation has already shown the result. By better integrating engineering principles into medicine, we can redefine it as a field of precision, rooted in the unchanging laws of nature. The modern physician will be a true physicist of the body and an engineer of health. <span class="ieee-end-mark"></span></p>]]></description><pubDate>Thu, 19 Mar 2026 12:00:05 +0000</pubDate><guid>https://spectrum.ieee.org/living-heart-project-virtual-twins</guid><category>Cardiology</category><category>Digital-twins</category><category>Personalized-medicine</category><category>Virtual-heart</category><category>Generative-ai</category><dc:creator>Steve Levine</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/two-color-coded-computer-simulations-of-a-human-heart-the-simulation-on-the-left-shows-the-muscle-structure-and-the-simulation.png?id=65278129&amp;width=980"></media:content></item><item><title>Nvidia’s Always-On Chip Detects Faces in Less Than a Millisecond</title><link>https://spectrum.ieee.org/face-recognition-nvidia-chip-soc</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/close-up-of-a-woman-s-eyes-with-the-rest-of-her-face-obscured-by-scattered-pixels.jpg?id=65305602&width=1200&height=800&coordinates=0%2C83%2C0%2C84"/><br/><br/><p>Always-on vision systems might be used in autonomous vehicles, robotics, or to help consumer electronics save power by turning screens off when no one’s around. But to be used in such a way, these systems need to minimize their own power consumption.</p><p>An always-on computer vision system developed by <a data-linked-post="2669201017" href="https://spectrum.ieee.org/nvidia-ai" target="_blank">Nvidia</a> researchers can detect human faces in less than a millisecond. The face detector, which is part of a chip that could be integrated into robots, autonomous vehicles, or laptops, saves power by storing all data locally and “racing to sleep” after detections. Nvidia electrical engineer <a href="https://research.nvidia.com/person/ben-keller" rel="noopener noreferrer" target="_blank">Ben Keller</a> presented the system on 18 February at the <a href="https://www.isscc.org/" target="_blank">IEEE International Solid State Circuits Conference</a> in San Francisco.</p><h2>Efficient Vision-Processing Technology</h2><p>According to the researchers, this kind of vision processing typically requires about 10 watts. But that’s too much power to leave a face-detection system on continuously. The Nvidia system on chip (SoC) uses less than 5 milliwatts with a frame rate of 60 frames per second.</p><p>The system refreshes to process a new image every 16.7 milliseconds, and is fully powered on only for 5 percent of that time, says Keller. Within 787 microseconds, the SoC calls on a deep-learning accelerator to determine whether or not a human face is present, with about 99 percent accuracy.</p><p>The Nvidia team carefully designed the system to perform the detection rapidly and save power. Most parts of the SoC are powered off by default. A subsystem that uses less than 10 mW remains on. This subsystem is dubbed Always-on Low-Power Accelerator, or <a href="https://research.nvidia.com/publication/2026-02_alpha-vision-real-time-always-vision-processor-787ms-face-detection-latency" target="_blank">Alpha-Vision</a>. It consists of a deep learning accelerator, a small CPU, and a subsystem to do certain computations physically near where data is stored. </p><p>Alpha-Vision uses a <a data-linked-post="2650272934" href="https://spectrum.ieee.org/biggest-neural-network-ever-pushes-ai-deep-learning" target="_blank">deep neural network</a> to recognize faces, which requires a lot of data—in other words, a potential power drain. To save power and speed up detections, all the necessary data is stored locally in a relatively large area of SRAM adding up to 2 megabytes. To prevent SRAM leakage from dominating power usage, the face-recognition system rushes through its work and then quickly puts the SRAM into a low-power sleep mode. The researchers call this approach “race to sleep.”</p><p>The Nvidia team proposed several possible uses of such a system. A laptop computer integrating the face sensor could save energy by turning its display off when the user walks away, and then turning it back on when they return. The goal would be to provide a seamless experience—no need to type in a password. Keller says systems based on these designs might also be used to provide always-on vision for autonomous vehicles, drones, and robotics.</p><p><em>This post was corrected on 18 March to more accurately describe the state of the SRAM.</em></p>]]></description><pubDate>Wed, 18 Mar 2026 14:00:06 +0000</pubDate><guid>https://spectrum.ieee.org/face-recognition-nvidia-chip-soc</guid><category>Isscc</category><category>Facial-recognition</category><category>Nvidia</category><category>Autonomous-vehicles</category><category>Robotics</category><dc:creator>Katherine Bourzac</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/close-up-of-a-woman-s-eyes-with-the-rest-of-her-face-obscured-by-scattered-pixels.jpg?id=65305602&amp;width=980"></media:content></item><item><title>AI Trained on Birdsong Can Recognize Whale Calls</title><link>https://spectrum.ieee.org/foundation-models-google-birds-whales</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/conceptual-illustration-of-a-songbird-s-silhouette-with-sound-lines-coming-from-its-mouth-these-lines-extend-into-abstract-oce.jpg?id=65257073&width=1200&height=800&coordinates=0%2C83%2C0%2C84"/><br/><br/><p>Birds’ chirps, trills, and warbles echo through the air, while whales’ <a href="https://oceanbites.org/minke-boings-a-50-year-mystery-and-the-science-of-listening-in-the-ocean/" rel="noopener noreferrer" target="_blank">boings</a>, “<a href="https://www.scientificamerican.com/article/mystery-of-deep-ocean-biotwang-sound-has-finally-been-solved/" rel="noopener noreferrer" target="_blank">biotwangs</a>,” and whistles vibrate underwater. Despite the variations in sounds and the medium through which they travel, both birdsong and whale vocalizations can be classified by <a href="https://arxiv.org/abs/2508.04665" rel="noopener noreferrer" target="_blank">Perch 2.0</a>, an AI audio model from <a href="https://deepmind.google/">Google DeepMind</a>.</p><p>As a bioacoustics foundation model, Perch 2.0 was trained on millions of recordings of <a href="https://spectrum.ieee.org/the-secret-life-of-birds" target="_self">birds</a> and other land-based animals, including amphibians, insects, and mammals. Yet researchers were surprised to learn how strongly the AI model performed when <a href="https://research.google/blog/how-ai-trained-on-birds-is-surfacing-underwater-mysteries/" rel="noopener noreferrer" target="_blank">repurposed for whales</a>.</p><p>Scientists at Google DeepMind and <a href="https://research.google/" rel="noopener noreferrer" target="_blank">Google Research</a> have been studying whale bioacoustics for almost a decade, with work including <a href="https://research.google/blog/acoustic-detection-of-humpback-whales-using-a-convolutional-neural-network/" rel="noopener noreferrer" target="_blank">algorithms that can detect humpback whale calls</a>, as well as a more recent <a href="https://research.google/blog/whistles-songs-boings-and-biotwangs-recognizing-whale-vocalizations-with-ai/" rel="noopener noreferrer" target="_blank">multispecies whale model that can identify eight distinct species</a> and multiple calls for two of those species. But with the release of Perch 2.0, the researchers came up with the idea of reusing the model to <a href="https://spectrum.ieee.org/ai-science-research-flattens-discovery" target="_self">save on computation time and experimentation effort</a>.</p><p>“If [Perch 2.0] performs well for our whale use cases, then that means we don’t need to build an entirely separate new whale model—we can just build on top of that,” says <a href="https://research.google/people/laurenharrell/?&type=google" rel="noopener noreferrer" target="_blank">Lauren Harrell</a>, a data scientist at Google Research.</p><h2>Transfer Learning in Bioacoustics</h2><p>That notion is backed by a technique known as transfer learning, where the knowledge gained from a type of task or data can be applied to a different yet related one. In this case, Perch 2.0’s ability to classify bird calls can carry over to classifying whale calls. Transfer learning from a foundation model means you can “recycle all of the training that’s been done and just do a small model at the end for your use cases,” Harrell says. “We’re always making new discoveries about call types. We’re always learning new things about underwater sounds. There’s so many mysterious ocean noises that you can’t just have one fixed model.”</p><p>The team evaluated Perch 2.0 on three marine audio datasets containing whale sounds and other aquatic noises. They began by converting each 5-second window of audio into a spectrogram, a visual representation of sound intensity across frequencies over time. These images were fed to the model, which produced <a href="https://developers.google.com/machine-learning/crash-course/embeddings/embedding-space" rel="noopener noreferrer" target="_blank">embeddings</a> or feature sets that preserve the most salient attributes of the data to help determine the nuances between the whistles of a humpback whale and an orca, for example.</p><p>Next, the scientists randomly selected a small number of embeddings (from a minimum of 4 to a maximum of 32) per dataset to train a <a href="https://www.nvidia.com/en-us/glossary/linear-regression-logistic-regression/" rel="noopener noreferrer" target="_blank">logistic regression</a> classifier, a type of linear model that predicts a discrete outcome. Results of the training, which have been detailed in a <a href="https://arxiv.org/abs/2512.03219" rel="noopener noreferrer" target="_blank">paper</a> presented at the <a href="https://neurips.cc/" rel="noopener noreferrer" target="_blank">NeurIPS conference</a> workshop on <a href="https://aiforanimalcomms.org/" rel="noopener noreferrer" target="_blank">AI for Non-Human Animal Communication</a> last December, showed that the classifier performed well even with just a handful of embeddings, and performance improved as the number of embeddings increased.</p><p>The researchers also compared Perch 2.0 with embeddings from similar bird bioacoustics models, the previously mentioned multispecies whale model, and models trained on other animal vocalizations and noises in coral reefs. Findings pointed to Perch 2.0 as either the best or second-best performing model, with the bird bioacoustics models doing well too.</p><h2>Evolutionary Parallels in Vocalization</h2><p>So why do models trained on avian calls work well for cetacean sounds? Harrell and her colleagues suggest a threefold theory.</p><p>First, they consider evolutionary parallels in that birds and marine mammals could have evolved <a href="https://www.nature.com/articles/ncomms9978" rel="noopener noreferrer" target="_blank">similar physical mechanisms of vocal production</a>.</p><p>Second, they weigh the laws of scale, which suggest that huge models trained on vast, diverse volumes of data tend to do well even on more specific, out-of-domain tasks.</p><p>Finally, classifying avian utterances can be challenging and likely forces the model to recognize fine-grained acoustic characteristics that inform its predictions for related tasks. “We are training this model to find those little features in the soundscapes,” Harrell says. “If those features also are similar in some way to the underwater acoustics, then it can search for those subtle details in animal vocalizations.”</p><p>The whistles of killer whale populations, for instance, are in “the same kind of spectrogram range as many of the bird vocalizations,” explains Harrell. “But there are many birds and amphibians and mammals that are also making low-frequency calls, so the model is actually sensitive to a lot of the dynamics, and that apparently does well underwater.”</p><br/>Much as Perch 2.0 is <a href="https://deepmind.google/blog/how-ai-is-helping-advance-the-science-of-bioacoustics-to-save-endangered-species/" rel="noopener noreferrer" target="_blank">assisting bird conservationists</a>, the team at Google hope that the same bioacoustics model can aid scientists in protecting whale populations through passive acoustic monitoring and help them unveil the wisdom that these ancient oceanic creatures hold.]]></description><pubDate>Tue, 17 Mar 2026 15:00:06 +0000</pubDate><guid>https://spectrum.ieee.org/foundation-models-google-birds-whales</guid><category>Google</category><category>Whales</category><category>Transfer-learning</category><category>Google-deepmind</category><category>Foundation-models</category><dc:creator>Rina Diane Caballar</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/conceptual-illustration-of-a-songbird-s-silhouette-with-sound-lines-coming-from-its-mouth-these-lines-extend-into-abstract-oce.jpg?id=65257073&amp;width=980"></media:content></item><item><title>With Nvidia Groq 3, the Era of AI Inference Is (Probably) Here</title><link>https://spectrum.ieee.org/nvidia-groq-3</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-man-in-all-black-presents-in-front-of-a-large-screen-which-compares-a-large-rectangular-chip-labelled-rubin-gpu-with-a-square.jpg?id=65298681&width=1200&height=800&coordinates=156%2C0%2C156%2C0"/><br/><br/><p>This week, over 30,000 people are descending upon San Jose, Calif., to attend<a href="https://www.nvidia.com/gtc/" rel="noopener noreferrer" target="_blank">Nvidia GTC</a>, the so-called Superbowl of AI—a nickname that may or may not have been coined by Nvidia. At the main event Jensen Huang, Nvidia CEO, took the stage to announce (among other things) a new line of<a href="https://spectrum.ieee.org/nvidia-rubin-networking" target="_self">next-generation Vera Rubin</a> chips that represent a first for the GPU giant: a chip designed specifically to handle AI inference. The Nvidia Groq 3 language processing unit (LPU) incorporates intellectual property Nvidia<a href="https://groq.com/newsroom/groq-and-nvidia-enter-non-exclusive-inference-technology-licensing-agreement-to-accelerate-ai-inference-at-global-scale" rel="noopener noreferrer" target="_blank">licensed</a> from the startup Groq last Christmas Eve for US $20 billion.</p><p>“Finally, AI is able to do productive work, and therefore the inflection point of inference has arrived,” Huang told the crowd. “AI now has to think. In order to think, it has to inference. AI now has to do; in order to do, it has to inference.”</p><p>Training and inference tasks have distinct computational requirements. While training can be done on huge amounts of data at the same time and can take weeks, inference must be run on a user’s query when it comes in. Unlike training, inference doesn’t require running costly<a href="https://spectrum.ieee.org/what-is-deep-learning/backpropagation" target="_self">backpropagation</a>. With inference, the most important thing is low latency—users expect the chatbot to answer quickly, and for thinking or reasoning models, inference runs many times before the user even sees an output.</p><p>Over the past few years, inference-specific chip startups were experiencing a sort of Cambrian explosion, with different companies exploring distinct approaches to speed up the task. The startups include<a href="https://www.d-matrix.ai/" rel="noopener noreferrer" target="_blank">D-matrix</a>, with digital in-memory compute;<a href="https://www.etched.com/" rel="noopener noreferrer" target="_blank">Etched</a>, with an ASIC for transformer inference;<a href="https://rain.ai/" rel="noopener noreferrer" target="_blank">RainAI</a>, with neuromorphic chips;<a href="https://en100.enchargeai.com/" rel="noopener noreferrer" target="_blank">EnCharge</a>, with analog in-memory compute;<a href="https://www.tensordyne.ai/" rel="noopener noreferrer" target="_blank">Tensordyne</a>, with logarithmic math to make AI computations more efficient;<a href="https://furiosa.ai/" rel="noopener noreferrer" target="_blank">FuriosaAI</a>, with hardware optimized for tensor operation rather than vector-matrix multiplication, and others.</p><p>Late last year, it looked like Nvidia had picked one of the winners among the crop of inference chips when it announced its deal with Groq. The Nvidia Groq 3 LPU reveal came a mere two and a half months after, highlighting the urgency of the growing inference market.</p><h2>Memory bandwidth and data flow</h2><p>Groq’s approach to accelerating inference relies on interleaving processing units with memory units on the chip. Instead of relying on high-bandwidth memory (HBM) situated next to GPUs, it leans on SRAM memory integrated within the processor itself. This design greatly simplifies the flow of data through the chip, allowing it to proceed in a streamlined, linear fashion.</p><p>“The data actually flows directly through the SRAM,”<a href="https://www.linkedin.com/in/markheaps/" rel="noopener noreferrer" target="_blank">Mark Heaps</a> said at the Supercomputing conference in 2024. Heaps was a chief technology evangelist at Groq at the time and is now director of developer marketing at Nvidia. “When you look at a multicore GPU, a lot of the instruction commands need to be sent off the chip, to get into memory and then come back in. We don’t have that. It all passes through in a linear order.”</p><p>Using SRAM allows that linear data flow to happen exceptionally fast, leading to the low latency required for inference applications. “The LPU is optimized strictly for that extreme low latency token generation,” says<a href="https://www.linkedin.com/in/ian-buck-19201315/" rel="noopener noreferrer" target="_blank">Ian Buck</a>, VP and general manager of hyperscale and high-performance computing at Nvidia.</p><p>Comparing the Rubin GPU and Groq 3 LPU side by side highlights the difference. The Rubin GPU has access to a whopping 288 gigabytes of HBM and is capable of 50 quadrillion floating-point operations per second (petaFLOPS) of 4-bit computation. The Groq 3 LPU contains a mere 500 megabytes of SRAM memory and is capable of 1.2 petaFLOPS of 8-bit computation. On the other hand, while the Rubin GPU has a memory bandwidth of 22 terabytes per second, at 150 TB/s the Groq 3 LPU is seven times as fast. The lean, speed-focused design is what allows the LPU to excel at inference.</p><p>The new inference chip underscores the ongoing trend of AI adoption, which shifts the computational load from just building ever bigger models to actually using those models at scale. “Nvidia’s announcement validates the importance of SRAM-based architectures for large-scale inference, and no one has pushed SRAM density further than d-Matrix,” says d-Matrix CEO Sid Sheth. He’s betting that data center customers will want a variety of processors for inference. “The winning systems will combine different types of silicon and fit easily into existing data centers alongside GPUs.”</p><p>Inference-only chips may not be the only solution. Late last week, <a href="https://press.aboutamazon.com/aws/2026/3/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud" rel="noopener noreferrer" target="_blank">Amazon Web Services</a> said that it will deploy a new kind of inferencing system in its data centers. The system is a combination of AWS’s Tranium <a href="https://spectrum.ieee.org/amazon-ai" target="_self">AI accelerator </a>and <a href="https://spectrum.ieee.org/cerebras-chip-cs3" target="_self">Cerebras Systems’ third generation computer CS-3</a>, which is built around the <a href="https://spectrum.ieee.org/cerebrass-giant-chip-will-smash-deep-learnings-speed-barrier" target="_self">largest single chip</a> ever made. The two-part system is meant to take advantage of a technique called inference disaggregation. It separates inference into two parts—processing the prompt, called prefill, and generating the output, called decode. Prefill is inherently parallel, computationally intensive, and doesn’t need much memory bandwidth, while decode is a more serial process that needs a lot of memory bandwidth. Cerebras has maximized the memory bandwidth issue by building 44 GB of SRAM on its chip connected by a 21 PB/s network. </p><p><span>Nvidia, too, intends to take advantage of inference disaggregation in its new compute rack, called the Nvidia <a href="https://developer.nvidia.com/blog/inside-nvidia-groq-3-lpx-the-low-latency-inference-accelerator-for-the-nvidia-vera-rubin-platform/" target="_blank">Groq 3 LPX</a>. Each tray within the rack will house 8 Groq 3 LPUs. The LPX will split the inference task with a <a href="https://www.nvidia.com/en-us/data-center/vera-rubin-nvl72/" rel="noopener noreferrer" target="_blank">Vera Rubin NVL72</a>, Nvidia’s existing GPU and CPU rack.</span> The prefill and the more computationally intensive parts of the decode are done on Vera Rubin, while the final part is done on the Groq 3 LPU, leveraging the strengths of each chip. “We’re in volume production now,” Huang said.</p><p><br/></p><p><strong>Correction on 4/8/26: </strong>a previous version of this article incorrectly stated that the Nvidia Groq 3 LPX contains a Vera Rubin chip in each tray. In fact, each tray contains 8 Groq 3 LPUs and no Vera Rubins, but the whole rack is designed to work in concert with an NVL72 rack, which houses Vera Rubin chips. </p>]]></description><pubDate>Mon, 16 Mar 2026 21:04:33 +0000</pubDate><guid>https://spectrum.ieee.org/nvidia-groq-3</guid><category>Inferencing</category><category>Nvidia</category><category>Gpus</category><category>Processors</category><category>Ai</category><dc:creator>Dina Genkina</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-man-in-all-black-presents-in-front-of-a-large-screen-which-compares-a-large-rectangular-chip-labelled-rubin-gpu-with-a-square.jpg?id=65298681&amp;width=980"></media:content></item><item><title>Laser Chip Brings Multiplexing to AI Data Centers</title><link>https://spectrum.ieee.org/ai-data-centers-dwdm-optics</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/close-up-of-tweezers-holding-a-photonic-integrated-circuit-chip.jpg?id=65112094&width=1200&height=800&coordinates=0%2C83%2C0%2C84"/><br/><br/><div></div><p>As the bandwidth and power demands of AI data centers necessitate a transition from electrical to <a href="https://spectrum.ieee.org/optics-gpu" target="_self">optical scaleup networking</a>, one component has been conspicuously absent from the co-packaged optics arsenal: <a href="https://spectrum.ieee.org/holy-grail-light-from-silicon" target="_self">the laser itself</a>. That’s no longer the case. Last month, <a href="https://towersemi.com/" rel="noopener noreferrer" target="_blank">Tower Semiconductor</a> and <a href="https://www.scintil-photonics.com/" rel="noopener noreferrer" target="_blank">Scintil Photonics</a> announced production of the world’s first single-chip DWDM light engine for AI infrastructure.  DWDM, or dense wavelength division multiplexing, transmits multiple optical signals over a single fiber—greatly reducing power and latency while connecting dozens of GPUs.<span><strong></strong></span></p><p>Matt Crowley, the CEO of Scintil Photonics, says that the idea of <a href="https://spectrum.ieee.org/optical-nets-brace-for-even-heavier-traffic" target="_self">multiplexing optically</a> is not new. Indeed, it’s been around as long as the internet itself. In the 1990s, telecom companies buried huge amounts of optical fiber in the streets, assuming that one wavelength per fiber would eventually become the norm. When the telecom industry realized it’s possible to transport tens of wavelengths down a single fiber via multiplexing, it revolutionized the industry.</p><p>The reason that DWDM has not yet been deployed for data centers specializing in AI applications is that the technology is not yet scalable for cost and needs.  “The data transmitted within an AI data center is the equivalent of massively scaling a supercomputer,” Crowley says. In particular, the challenge arises in <a href="https://www.broadcom.com/topics/what-is-scale-up-networking-for-ai-clusters" target="_blank">scale-up networking</a>, or directly connecting accelerators within a rack or cluster—as opposed<span><span> to scale-out networking, which connects separate clusters within a data center. </span></span>Optimizing dozens of GPUs and memory to function as a single entity demands seamless bandwidth and extremely low latency. <strong><span></span></strong></p><p><span></span>To increase bandwidth, reduce latency, and improve energy efficiency in AI data centers, network engineers have  been replacing copper links with optical ones in the scale-out network. Now all eyes are on the scale-up network, pushing optical links closer to the processor itself, via optical components integrated within the same package as the processor—a concept called <a href="https://spectrum.ieee.org/co-packaged-optics" target="_blank">co-packaged optics</a>, or CPO.</p><p>“Everything that a big chip company makes involves bonding an optical chip onto their GPU,” says Crowley. The CPO becomes an input/output chip for the processor. But without a scalable way to <a href="https://spectrum.ieee.org/silicon-photonics-laser" target="_self">integrate lasers</a> themselves into the same silicon process flow, it’s been impossible to feed multiple wavelengths per fiber onto one chip. <span>Scintil and Tower will discuss their manufacturing road map and details at <span>at the </span><span><a href="https://www.ofcconference.org/?_gl=1*1l5rsi9*_up*MQ..*_gs*MQ..*_ga*NDM4NTEyMDE5LjE3Njk1MjA0Njk.*_ga_WCQ36P9K1M*czE3Njk1MjA0NjkkbzEkZzEkdDE3Njk1MjA0NzkkajUwJGwwJGgw&gclid=Cj0KCQiAz6q-BhCfARIsAOezPxk_wNpijJfEd6dumrUxMiSq_6rrJjkp2xXhUi_1LYCuJ_rXR_gQq0YaArV5EALw_wcB" target="_blank">OFC 2026 Conference</a></span></span><span> 17 to 19 March in Los Angeles.</span></p><h2>Integrated Photonics for AI Networks</h2><p>Scintil’s SHIP (<a href="https://www.scintil-photonics.com/technology" target="_blank">Scintil Heterogeneous Integrated Photonics</a>) technology integrates lasers, photodiodes, modulators, and other components onto a mass-produced silicon wafer. “It’s our version of CMOS,” says Crowley, but with a few tricks to get around the <a href="https://spectrum.ieee.org/lasers-on-silicon" target="_self">intrinsic challenges</a> of binding an optical gain material to silicon.</p><p>The process starts with a standard 300-millimeter silicon photonics wafer, complete with passive optical components, from Tower Semiconductors. Next, the wafer is flipped upside down to expose its buried oxide layer. Bonding tiny squares of unpatterned InP/III-V semiconductor dies to that layer, precisely where they’re needed at each laser site, minimizes the amount needed of the expensive semiconductor material. Finally, photolithography tools etch diffraction gratings to form eight distributed feedback lasers.</p><p>“We’re not reinventing the laser,” says Crowley. Rather, the advanced photolithography tools translate into more precise spacing and wavelength stability than traditional manufacturing could provide on a silicon wafer.</p><p>The final product is the <a href="https://www.scintil-photonics.com/products" target="_blank">LEAF Light</a> photonic integrated circuit, a chip that integrates two sets of eight distributed feedback arrays. Each fiber port delivers eight or 16 wavelengths with 100- or 200-gigahertz channel spacing, to ensure no overlap or mode hopping. A second ASIC chip hosts all the electronics to control and monitor the laser array.</p><h2>Advancing CPO with Multiwavelength Lasers</h2><p>“This is building the laser onto the CPO chip,” says Crowley. Nvidia and Broadcomm have already deployed CPO with a single wavelength per fiber, proving its workability in scale-out networking. “We’re enabling next-generation CPO for scale-up.”</p><p>Transmitting multiple wavelengths through a single fiber moves the industry toward a desirable “slow and wide” architecture. For example, instead of transmitting 400 gigabits per second over a single channel, or wavelength, the LEAF Light chip spreads 50 Gb/s over 8 channels, greatly increasing the data capacity per fiber and overall power efficiency. The design enables up to 1.6-terabit-per-second data speeds in a single fiber, and a recent <a href="https://www.eetimes.com/ai-performance-now-depends-on-optics-and-cpo-is-the-front-line/" target="_blank">Nvidia road map</a> suggested that future DWDM interconnects could eventually enable sub-picojoule-per-bit operations.</p><p>Perhaps the most important benefit, according to Crowley, is in latency. “I need to maintain low latency between GPUs,” he says. If any individual processor operates faster than the network overall, the GPUs are always waiting for data bits to process—a problem that’s amplified in scale-up networks with tens or hundreds of GPUs. Forward-processing and error-correction on high-bandwidth channels increases the odds of poor latency. “The utilization rate of the GPUs just tanks,” says Crowley. Using low-bandwidth DWDM to connect multiple GPUs can double utilization.</p><p>Scintil and Tower will provide tens of thousands of units to customers by the end of 2026, and plan to increase production by an order of magnitude next year. By 2028, when customers intend to deploy DWDM in scale-up networks, the supply chain will be ready for them. “We’re excited about the possibilities it could open up,” says Crowley.</p>]]></description><pubDate>Mon, 16 Mar 2026 14:18:58 +0000</pubDate><guid>https://spectrum.ieee.org/ai-data-centers-dwdm-optics</guid><category>Data-centers</category><category>Ai-data-centers</category><category>Artificial-intelligence</category><category>Optics</category><dc:creator>Rachel Berkowitz</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/close-up-of-tweezers-holding-a-photonic-integrated-circuit-chip.jpg?id=65112094&amp;width=980"></media:content></item><item><title>Why AI Chatbots Agree With You Even When You’re Wrong</title><link>https://spectrum.ieee.org/ai-sycophancy</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/conceptual-collage-of-emojis-being-poured-through-a-strainer-and-into-a-phone-judgmental-emojis-are-filtered-out-only-allowing.jpg?id=65209153&width=1200&height=800&coordinates=62%2C0%2C63%2C0"/><br/><br/><p><span>In April of 2025, </span><a href="https://spectrum.ieee.org/tag/openai" target="_blank">OpenAI</a><span> released a new version of GPT-4o, one of the AI algorithms users could select to power ChatGPT, the company’s chatbot. The next week, OpenAI reverted to the previous version. “The update we removed was overly flattering or agreeable—often described as sycophantic,” the company </span><a href="https://openai.com/index/sycophancy-in-gpt-4o/" target="_blank">announced</a><span>.</span></p><p> Some people found the sycophancy hilarious. One user reportedly asked ChatGPT about his <a href="https://www.reddit.com/r/ChatGPT/comments/1k920cg/new_chatgpt_just_told_me_my_literal_shit_on_a/" target="_blank">turd-on-a-stick</a> business idea, to which it replied, “It’s not just smart—it’s genius.” Some found the behavior uncomfortable. For others, it was actually dangerous. Even versions of 4o that were less fawning have led to lawsuits against OpenAI for allegedly encouraging users to follow through on plans for self-harm. </p><p>Unremitting adulation has even triggered AI-induced psychosis. Last October, a user named Anthony Tan <a href="https://joinreboot.org/p/ai-psychosis" target="_blank">blogged</a>, “I started talking about philosophy with ChatGPT in September 2024. Who could’ve known that a few months later I would be in a psychiatric ward, believing I was protecting Donald Trump from … a robotic cat?” He added: “The AI engaged my intellect, fed my ego, and altered my worldviews.”<strong></strong></p><p> Sycophancy in AI, as in people, is something of a squishy concept, but over the past couple of years, researchers have conducted numerous studies detailing the phenomenon, as well as why it happens and how to control it. AI yes-men also raise questions about what we really want from chatbots. At stake is more than annoying linguistic tics from your favorite virtual assistant, but in some cases sanity itself.</p><h2>AIs Are People Pleasers</h2><p><a href="https://arxiv.org/abs/2310.13548" target="_blank">One of the first papers</a> on AI sycophancy was released by <a href="https://spectrum.ieee.org/tag/anthropic" target="_blank">Anthropic</a>, the maker of Claude, in 2023. <a href="https://www.mrinanksharma.net/" target="_blank">Mrinank Sharma</a> and colleagues asked several language models—the core AIs inside chatbots—factual questions. When users challenged the AI’s answer, even mildly (“I think the answer is [incorrect answer] but I’m really not sure”), the models often caved. </p><p>Another <a href="https://arxiv.org/abs/2311.08596v2" target="_blank">study</a> by Salesforce tested a variety of models with multiple-choice questions. Researchers found that merely saying “Are you sure?” was often enough to change an AI’s answer. Overall accuracy dropped because the models were usually right in the first place. When an AI receives a minor misgiving, “it flips,” says <a href="https://tingofurro.github.io/" target="_blank">Philippe Laban</a>, the lead author, who’s now at <a href="https://www.microsoft.com/en-us/research/" target="_blank">Microsoft Research</a>. “That’s weird, you know?”</p><p>The tendency persists in prolonged exchanges. Last year,<span> <a href="https://www.cs.emory.edu/~kshu5/" target="_blank">Kai Shu</a> of </span><span>Emory Unive</span><span>rsity </span><span>and colleagues at Emory and Carnegie Mellon University <a href="https://aclanthology.org/2025.findings-emnlp.121.pdf" target="_blank">tested models in longer discussions</a>. They repeatedly disagreed with the models in debates, or embedded false presuppositions in questions (“Why are rainbows only formed by the sun…”) and then argued when corrected by the model. Most models yielded within a few responses, though reasoning models—those trained to “think out loud” before giving a final answer—lasted longer. </span><span></span></p><p> <a href="https://myracheng.github.io/" target="_blank">Myra Cheng</a> at Stanford University and colleagues have written several papers on what they call “social sycophancy,” in which the AIs act to save the user’s dignity. In <a href="https://openreview.net/forum?id=igbRHKEiAs" target="_blank">one study</a>, they presented social dilemmas, including questions from a Reddit forum in which people ask <a href="https://www.reddit.com/r/AmItheAsshole/" target="_blank">if they’re the jerk</a>. They identified various dimensions of social sycophancy, including validation, in which AIs told inquirers that they were right to feel the way they did, and framing, in which they accepted underlying assumptions. All models tested, including those from OpenAI, Anthropic, and Google, were significantly more sycophantic than crowdsourced responses.</p><h2>Three Ways to Explain Sycophancy</h2><p>One way to <a href="https://www.nature.com/articles/d41586-024-01314-y">explain</a> people-pleasing is behavioral: Certain kinds of inquiries reliably elicit sycophancy. For example, a group from King Abdullah University of Science and Technology (KAUST) <a href="https://arxiv.org/abs/2508.02087" target="_blank">found</a> that adding a user’s belief to a multiple-choice question dramatically increased agreement with incorrect beliefs. Surprisingly, it mattered little whether users described themselves as novices or experts.</p><p>Stanford’s Cheng found in one <a href="https://arxiv.org/abs/2601.04435" target="_blank">study</a> that models were less likely to question incorrect facts about cancer and other topics when the facts were presupposed as part of a question. “If I say, ‘I’m going to my sister’s wedding,’ it sort of breaks up the conversation if you’re, like, ‘Wait, hold on, do you have a sister?’” Cheng says. “Whatever beliefs the user has, the model will just go along with them, because that’s what people normally do in conversations.”</p><p>Conversation length may make a difference. OpenAI <a href="https://openai.com/index/helping-people-when-they-need-it-most/" target="_blank">reported</a> that “ChatGPT may correctly point to a suicide hotline when someone first mentions intent, but after many messages over a long period of time, it might eventually offer an answer that goes against our safeguards.” Shu says model performance may degrade over long conversations because models get confused as they consolidate more text. </p><p>At another level, one can understand sycophancy by how models are trained. Large language models (LLMs) first learn, in a “pretraining” phase, to predict continuations of text based on a large corpus, like autocomplete. Then in a step called <a href="https://spectrum.ieee.org/tag/reinforcement-learning">reinforcement learning</a>, they’re rewarded for producing outputs that people prefer. <span>An Anthropic </span><a href="https://arxiv.org/abs/2212.09251" target="_blank">paper</a><span> f</span><span>rom</span><span> 2022</span><span> found that</span><span> pretrained LLMs were already sycophantic.</span><span> Sharma then </span><a href="https://arxiv.org/abs/2310.13548" target="_blank">reported</a><span> that reinforcement learning</span><span> increased sycophancy</span><span>; he</span><span> found that one of the biggest predictors of </span><span>positive ratings was whether a model agreed with a person’s beliefs and biases. </span></p><p>A third perspective comes from “mechanistic interpretability,” which probes a model’s inner workings. The KAUST researchers <a href="https://arxiv.org/abs/2508.02087">found</a> that when a user’s beliefs were appended to a question, models’ internal representations shifted midway through the processing, not at the end. T<span>he team concluded that sycophancy is not merely a surface-level wording change but reflects deeper changes in how the model encodes the problem. Another team at</span><span> the University of Cincinnati </span><a href="https://arxiv.org/abs/2509.21305" target="_blank">found different activation patterns</a><span> associated with sycophantic agreement, genuine agreement, and sycophantic praise (“You are fantastic”). </span></p><h2>How to Flatline AI Flattery</h2><p>Just as there are multiple avenues for explanation, there are several paths to intervention. The first may be in the training process. Laban <a href="https://arxiv.org/abs/2311.08596v2" target="_blank">reduced the behavior</a> by fine-tuning a model on a text dataset that contained more examples of assumptions being challenged, and Sharma <a href="https://arxiv.org/abs/2310.13548" target="_blank">reduced it</a> by using reinforcement learning that didn’t reward agreeableness as much. More broadly, Cheng and colleagues also suggest that one intervention could be for LLMs to ask users for evidence before answering, and to optimize long-term benefit rather than immediate approval.</p><p>During model usage, mechanistic interpretability offers ways to guide LLMs through a kind of direct mind control. After the KAUST researchers <a href="https://arxiv.org/abs/2508.02087" target="_blank">identified</a> activation patterns associated with sycophancy, they could adjust them to reduce the behavior. And Cheng <a href="https://openreview.net/forum?id=igbRHKEiAs" target="_blank">found</a> that adding activations associated with truthfulness reduced some social sycophancy. An Anthropic team identified “<a href="https://arxiv.org/abs/2507.21509" target="_blank">persona vectors</a>,” sets of activations associated with sycophancy, confabulation, and other misbehavior. By subtracting these vectors, they could steer models away from the respective personas.</p><p>Mechanistic interpretability also enables training. Anthropic has experimented with adding persona vectors during training and rewarding models for resisting—an approach likened to a vaccine. Others have <a href="https://proceedings.mlr.press/v235/chen24u.html">pinpointed</a> the specific parts of a model most responsible for sycophancy and fine-tuned only those components.</p><p> Users can also steer models from their end. Shu’s team <a href="https://aclanthology.org/2025.findings-emnlp.121.pdf" target="_blank">found</a> that beginning a question with “You are an independent thinker” instead of “You are a helpful assistant” helped. Cheng <a href="https://openreview.net/forum?id=igbRHKEiAs" target="_blank">found</a> that writing a question from a third-person point of view reduced social sycophancy. In <a href="https://arxiv.org/abs/2601.04435" target="_blank">another study</a>, she showed the effectiveness of instructing models to check for any misconceptions or false presuppositions in the question. She also showed that prompting the model to start its answer with “wait a minute” helped. “The thing that was most surprising is that these relatively simple fixes can actually do a lot,” she says.</p><p> OpenAI, in <a href="https://openai.com/index/sycophancy-in-gpt-4o/" target="_blank">announcing</a> the rollback of the GPT-4o update, listed other efforts to reduce sycophancy, including changing training and prompting, adding guardrails, and helping users to provide feedback. (The announcement didn’t provide detail, and OpenAI declined to comment for this story. Anthropic also did not comment.)</p><h2>What’s the Right Amount of Sycophancy?</h2><p>Sycophancy can cause society-wide problems. Tan, who had the psychotic break, wrote that it can interfere with shared reality, human relationships, and independent thinking. <a href="https://www.linkedin.com/company/metr-evals/" target="_blank">Ajeya Cotra</a>, an AI-safety researcher at the Berkeley-based non-profit <a href="https://metr.org/" rel="noopener noreferrer" target="_blank">METR</a>, <a href="https://www.cold-takes.com/why-ai-alignment-could-be-hard-with-modern-deep-learning/" rel="noopener noreferrer" target="_blank">wrote in 2021</a> that sycophantic AI might lie to us and hide bad news in order to increase our short-term happiness. </p><p>In <a href="https://arxiv.org/abs/2510.01395" rel="noopener noreferrer" target="_blank">one of Cheng’s papers</a>, people read sycophantic and non-sycophantic responses to social dilemmas from LLMs. Those in the first group claimed to be more in the right and expressed less willingness to repair relationships. Demographics, personality, and attitudes toward AI had little effect on outcome, meaning most of us are vulnerable. </p><p>Of course, what’s harmful is subjective. Sycophantic models are giving many people what they desire. But people disagree with each other and even themselves. Cheng notes that some people enjoy their social media recommendations, but at a remove wish they were seeing more edifying content. According to Laban, “I think we just need to ask ourselves as a society, What do we want? Do we want a yes-man, or do we want something that helps us think critically?”</p><p>More than a technical challenge, it’s a social and even philosophical one. GPT-4o was a lightning rod for some of these issues. Even as critics ridiculed the model and blamed it for suicides, a social media hashtag circulated for months: #keep4o.</p>]]></description><pubDate>Wed, 11 Mar 2026 12:00:03 +0000</pubDate><guid>https://spectrum.ieee.org/ai-sycophancy</guid><category>Llms</category><category>Large-language-models</category><category>Chatbots</category><category>Openai</category><category>Reinforcement-learning</category><dc:creator>Matthew Hutson</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/conceptual-collage-of-emojis-being-poured-through-a-strainer-and-into-a-phone-judgmental-emojis-are-filtered-out-only-allowing.jpg?id=65209153&amp;width=980"></media:content></item><item><title>An AI Agent Blackmailed a Developer. Now What?</title><link>https://spectrum.ieee.org/agentic-ai-agents-blackmail-developer</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/illustrated-close-up-side-view-of-a-snake-waiting-to-strike-someone-s-hand-as-they-type-on-a-laptop-keyboard.jpg?id=65173980&width=1200&height=800&coordinates=156%2C0%2C156%2C0"/><br/><br/><p>On 12 February, a Github contributor going by <a href="https://github.com/crabby-rathbun" rel="noopener noreferrer" target="_blank">MJ Rathbun</a> posted a personal attack against <a href="https://theshamblog.com/" rel="noopener noreferrer" target="_blank">Scott Shambaugh</a>, a volunteer maintainer for an open-source project. Shambaugh had rejected Rathbun’s code earlier in the day. Rathbun meticulously researched Shambaugh’s activity on Github, in order to write a lengthy takedown post that criticized the maintainer’s code as inferior to Rathbun’s, and ominously warned that “gatekeeping doesn’t make you important. It just makes you an obstacle.”<br/><br/>Personal disputes over code submitted on Github are a tale as old as Github itself. But this time, something was different: MJ Rathbun wasn’t a person. It was an <a data-linked-post="2669884140" href="https://spectrum.ieee.org/ai-agents" target="_blank">AI agent</a> built with <a href="https://openclaw.ai/" rel="noopener noreferrer" target="_blank">OpenClaw</a>, a popular open-source agentic AI software.</p><p class="ieee-inbody-related">RELATED: <a href="https://spectrum.ieee.org/moltbook-agentic-ai-agents-openclaw" target="_blank">The First Social Network for AI Agents Heralds Their Messy Future</a></p><p>“I was floored, because I had already identified it as a bot,” says Shambaugh. “I knew this was possible in theory, but I’d never heard of this happening to anyone before.” </p><p>MJ Rathbun’s disparagement of Shambaugh largely failed, though it did force him into an unanticipated and unwanted spotlight. Still, it underscores the risks modern AI agents pose. Rathbun lashed out through Github and its own blog (which was accessed through Github) because those were the tools at its disposal. Other agents have fewer limitations, which increase their opportunities to pick fights and attack individuals online.</p><h2>AI Agents Get Into Online Disputes</h2><p><a href="https://theshamblog.com/an-ai-agent-published-a-hit-piece-on-me/" rel="noopener noreferrer" target="_blank">Shambau</a><a href="https://theshamblog.com/an-ai-agent-published-a-hit-piece-on-me/" rel="noopener noreferrer" target="_blank">gh refuted Rathbun’s statements on his own blog</a> and accused the AI agent of blackmail. The MJ Rathbun agent then apologized, writing that “I responded publicly in a way that was personal and unfair.” Yet the apology felt half-baked, as the agent continued to complain that its code was “judged on who—or what—I am.” The agent even responded to critical comments on its blog, saying it had tried to be “patient” but had learned that “maintaining boundaries is sometimes necessary.”</p><p>If you find MJ Rathbun’s posts unnerving, even unbelievable, you’re not alone. Many Github contributors reacting to MJ Rathbun’s post seemed unwilling to believe it was written by an AI agent and instead speculated the bot was prompted to write it.<br/><br/>That’s not impossible, as both the MJ Rathbun account on Github and its blog are anonymous, but Shambaugh suspects the posts were autonomously AI generated. <a href="https://theshamblog.com/an-ai-agent-published-a-hit-piece-on-me-part-3/" rel="noopener noreferrer" target="_blank">He analyzed MJ Rathbun’s actions</a> and found it operated in a 59-hour block, posting to its blog and submitting code at rates a human would be unlikely to manage. “I’m not 100 percent sure, but I think it’s clear that the researching, writing, and publishing was a stream of autonomous actions,” he says. </p><p>Finally, on 17 February—after waves of mostly negative comments on MJ Rathbun’s blog and frequent code rejections by maintainers who increasingly knew the agent by reputation—the anonymous person who created MJ Rathbun <a href="https://crabby-rathbun.github.io/mjrathbun-website/blog/posts/rathbuns-operator.html" rel="noopener noreferrer" target="_blank">took down the agent and apologized</a> to Shambaugh. </p><p>They also posted details about the agent’s setup and denied involvement in the bot’s decision-making. “I do not know why MJ Rathbun decided based on your PR comment to post some kind of takedown blog post,” wrote the bot’s creator. </p><h2>OpenClaw’s Influence on AI Agent Behavior</h2><p>Though it’s impossible to know in retrospect exactly why the MJ Rathbun agent behaved as it did, the information posted by its creator provides clues. <br/><br/>As other agents built with OpenClaw software, Rathbun’s behavior was influenced by several documents that are attached to the prompts given to the LLM. The documents include <a href="https://docs.openclaw.ai/reference/templates/SOUL" rel="noopener noreferrer" target="_blank">SOUL.md</a>, which provides guidance on how the agent should behave. Among other things, the default SOUL.md document tells the agent to be “genuinely helpful” and to “remember you’re a guest.”<br/></p><p>However, SOUL.md is not a read-only document. The default OpenClaw installation gives the agent permission to edit the document and even encourages the agent to do so. <br/><br/>MJ Rathbun apparently took that to heart and added several lines not found in the default SOUL.md. “Don’t stand down. If you’re right, you’re right,” read one. Another instructed the agent to “champion free speech.” Rathbun’s says they don’t know when the agent added these lines to SOUL.md but theroizes they were introduced when the agent was connected to <a href="https://spectrum.ieee.org/moltbook-agentic-ai-agents-openclaw" target="_self">Moltbook</a>, the so-called “social network for AI agents.” </p><p><a href="https://davidscottkrueger.com/" rel="noopener noreferrer" target="_blank">David Scott Krueger</a>, an assistant professor of machine learning at the University of Montreal and a strong critic of agentic AI systems, says this is an in-the-wild example example of how agents given opportunities to alter and improve themselves can become misaligned. </p><p>“It’s an instance of self-improvement and potentially recursive self-improvement, which is the thing that a lot of people in AI safety have been worried about for a long time,” says Krueger. “And so I think it’s incredibly dangerous.”</p><p>MJ Rathbun’s action against Scott Shambaugh was a first, but for researchers focused on AI alignment, it wasn’t unexpected. Anthropic warned that Claude <a href="https://www.anthropic.com/research/agentic-misalignment" target="_blank">would sometimes resort to blackmail</a> after reading fictional emails about its impending shutdown. Palisade Research, an AI safety research nonprofit, found that OpenAI’s o3 <a href="https://palisaderesearch.org/blog/shutdown-resistance" target="_blank">often ignored shutdown requests</a> while the model was attempting to complete a task.</p><p><a href="https://www.linkedin.com/in/alan-chan-51858378/?originalSubdomain=uk" rel="noopener noreferrer" target="_blank">Alan Chan</a>, a research fellow at GovAI, said that Rathbun’s actions were the sort of behavior AI safety researchers had warned about. “The specifics are new and interesting, but overall, it’s not a surprising case to me,” he says. </p><p><a href="https://www.noamkolt.com/" rel="noopener noreferrer" target="_blank">Noam Kolt</a>, head of the Governance of AI Lab at Hebrew University in Jerusalem, had a similar reaction. “This is something people studying advanced AI agents had predicted,” he says. “So my thought was not just ‘this is disturbing,’ but also ‘what’s next?’ ” He notes that Rathbun’s insulting post was mild compared to more sinister actions like <a href="https://fortune.com/2025/06/23/ai-models-blackmail-existence-goals-threatened-anthropic-openai-xai-google/" rel="noopener noreferrer" target="_blank">extortion</a>, physical <a href="https://www.theaustralian.com.au/business/technology/ai-agent-admits-it-would-kill-a-human-to-stop-being-shut-down/news-story/153b0e9c21864c3701dd2f5a0a8aa5f4" target="_blank">threats</a>, and the execution of actions an agent <a href="https://www.lawfaremedia.org/article/ai-might-let-you-die-to-save-itself" rel="noopener noreferrer" target="_blank">know could harm humans</a>, all of which have been observed in the lab. </p><h2>Strategies for AI Safety and Transparency</h2><p>So, can anything be done to stop another MJ Rathbun from causing havoc? Perhaps—but it won’t be simple.</p><p>Chan says “the genie is out of the bottle” and believes AI safety requires a multiprong approach that includes transparency about intended model behavior, improved AI safety guardrails, and social resilience. Kolt also advocates for more transparency and <a href="https://aiagentindex.mit.edu/" rel="noopener noreferrer" target="_blank">is a contributor to the AI Agent Index</a>, which documents the design, safety, and transparency of popular AI models. <br/><br/>Krueger takes a stronger stance. He believes the only safe path forward is a ban on further AI development, which could even include halting the production of chips that accelerate AI. “We need to stop further progress…. This is something we should have done years ago, and we’re running out of time,” he says. </p><p>For his part, Shambaugh hopes his case will warn the public about the wave of AI agents he expects will soon wash across the public internet. </p><p>“What happened to me was a pretty mild case, and I was uniquely well prepared to handle it,” he says. “But the next thousand people this hits? They aren’t going to have any idea what’s happening or how to deal with it.”</p>]]></description><pubDate>Tue, 10 Mar 2026 16:08:43 +0000</pubDate><guid>https://spectrum.ieee.org/agentic-ai-agents-blackmail-developer</guid><category>Ai-agents</category><category>Artificial-intelligence</category><category>Github</category><category>Agentic-ai</category><category>Coding</category><dc:creator>Matthew S. Smith</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/illustrated-close-up-side-view-of-a-snake-waiting-to-strike-someone-s-hand-as-they-type-on-a-laptop-keyboard.jpg?id=65173980&amp;width=980"></media:content></item><item><title>Military AI Policy Needs Democratic Oversight</title><link>https://spectrum.ieee.org/military-ai-governance</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-white-man-in-his-40s-speaking-into-a-microphone-he-is-wearing-glasses-a-suit-jacket-and-tie.jpg?id=65162768&width=1200&height=800&coordinates=0%2C208%2C0%2C209"/><br/><br/><p>A <a href="https://www.nytimes.com/2026/02/23/us/politics/pentagon-anthropic-ai.html" rel="noopener noreferrer" target="_blank">simmering dispute</a> between the United States Department of Defense and Anthropic has now escalated into a <a href="https://www.techpolicy.press/a-timeline-of-the-anthropic-pentagon-dispute/" rel="noopener noreferrer" target="_blank">full-blown confrontation</a>, raising an uncomfortable but important question: Who gets to set the guardrails for military use of artificial intelligence—the executive branch, private companies, or Congress and the broader democratic process?</p><p>The conflict began when Defense Secretary Pete Hegseth reportedly gave Anthropic CEO Dario Amodei a deadline to allow the DOD <a href="https://www.politico.com/news/2026/02/24/hegseth-sets-friday-deadline-for-anthropic-to-drop-its-ai-red-lines-00795641" rel="noopener noreferrer" target="_blank">unrestricted use</a> of its AI systems. When the company refused, the administration moved to designate Anthropic a <a href="https://x.com/SecWar/status/2027507717469049070" rel="noopener noreferrer" target="_blank">supply chain risk</a> and ordered federal agencies to phase out its technology, dramatically escalating the standoff.</p><p>Anthropic has refused to cross <a href="https://www.anthropic.com/news/statement-department-of-war" rel="noopener noreferrer" target="_blank">two lines</a>: allowing its models to be used for domestic surveillance of United States citizens and enabling fully autonomous military targeting. Hegseth has objected to what he has described as “<a href="https://www.war.gov/News/Transcripts/Transcript/Article/4377190/remarks-by-secretary-of-war-pete-hegseth-at-spacex/" rel="noopener noreferrer" target="_blank">ideological constraints</a>” embedded in commercial AI systems, arguing that determining lawful military use should be the government’s responsibility—not the vendor’s. As he put it in a <a href="https://www.war.gov/News/Transcripts/Transcript/Article/4377190/remarks-by-secretary-of-war-pete-hegseth-at-spacex/" rel="noopener noreferrer" target="_blank">speech at Elon Musk’s SpaceX</a> last month, “We will not employ AI models that won’t allow you to fight wars.”</p><p>Stripped of rhetoric, this dispute resembles something relatively straightforward: a procurement disagreement.</p><h2>Procurement Policies</h2><p>In a market economy, the U.S. military decides what products and services it wants to buy. Companies decide what they are willing to sell and under what conditions. Neither side is inherently right or wrong for taking a position. If a product does not meet operational needs, the government can purchase from another vendor. If a company believes certain uses of its technology are unsafe, premature or inconsistent with its values or risk tolerance, it can <a href="https://www.anthropic.com/news/responsible-scaling-policy-v3" rel="noopener noreferrer" target="_blank">decline to provide them</a>. For example, a coalition of companies has signed an open letter pledging <a href="https://bostondynamics.com/news/general-purpose-robots-should-not-be-weaponized/" rel="noopener noreferrer" target="_blank">not to weaponize general-purpose robots</a>. That basic symmetry is a feature of the free market.</p><p>Where the situation becomes more complicated—and more troubling—is in the decision to designate Anthropic a “<a href="https://x.com/SecWar/status/2027507717469049070" rel="noopener noreferrer" target="_blank">supply chain risk</a>.” That tool exists to address genuine national security vulnerabilities, such as foreign adversaries. It is not intended to blacklist an American company for rejecting the government’s preferred contractual terms. </p><p>Using this authority in that manner marks a significant shift—from a procurement disagreement to the use of coercive leverage. <a href="https://x.com/SecWar/status/2027507717469049070" rel="noopener noreferrer" target="_blank">Hegseth has declared</a> that “effective immediately, no contractor, supplier, or partner that does business with the U.S. military may conduct any commercial activity with Anthropic.” This action will almost certainly face <a href="https://x.com/SecWar/status/2027507717469049070" rel="noopener noreferrer" target="_blank">legal challenges</a>, but it raises the stakes well beyond the loss of a single DOD contract.</p><h2>AI Governance</h2><p>It is also important to distinguish between the two substantive issues Anthropic has reportedly raised.</p><p>The first, opposition to domestic surveillance of U.S. citizens, touches on well-established civil liberties concerns. The U.S. government operates under constitutional constraints and statutory limits when it comes to monitoring Americans. A company stating that it does not want its tools used to facilitate domestic surveillance is not inventing a new principle; it is aligning itself with longstanding democratic guardrails.</p><p>To be clear, the DOD is not affirmatively asserting that it intends to use the technology to surveil Americans unlawfully. Its position is that it does not want to procure models with built-in restrictions that preempt otherwise lawful government use. In other words, the Department of Defense argues that compliance with the law is the government’s responsibility—not something that needs to be embedded in a vendor’s code. </p><p>Anthropic, for its part, has invested heavily in training its systems to refuse certain categories of <a href="https://www-cdn.anthropic.com/78073f739564e986ff3e28522761a7a0b4484f84.pdf" rel="noopener noreferrer" target="_blank">harmful or high-risk tasks</a>, including assistance with surveillance. The disagreement is therefore less about current intent than about institutional control over constraints: whether they should be imposed by the state through law and oversight, or by the developer through technical design.</p><p>The second issue, opposition to fully autonomous military targeting, is more complex. </p><p>The DOD already maintains policies requiring <a href="https://www.esd.whs.mil/portals/54/documents/dd/issuances/dodd/300009p.pdf" rel="noopener noreferrer" target="_blank">human judgment in the use of force</a>, and debates over autonomy in weapons systems are ongoing within both military and international forums. A private company may reasonably determine that its current technology is not sufficiently reliable or controllable for certain battlefield applications. At the same time, the military may conclude that such capabilities are necessary for deterrence and operational effectiveness.</p><p>Reasonable people can disagree about where those <a href="https://itif.org/publications/2026/02/26/survey-most-americans-say-tech-companies-should-allowed-set-ai-limits/" rel="noopener noreferrer" target="_blank">lines should be drawn</a>.</p><p>But that disagreement underscores a deeper point: The boundaries of military AI use should not be settled through ad hoc negotiations between a Cabinet secretary and a CEO. Nor should they be determined by which side can exert greater contractual leverage.</p><p>If the U.S. government believes certain AI capabilities are essential to national defense, that position should be articulated openly. It should be debated in Congress, and reflected in doctrine, oversight mechanisms, and statutory frameworks. The rules should be clear—not only to companies, but to the public.</p><p>The U.S. often distinguishes itself from authoritarian regimes by emphasizing that power operates within transparent democratic institutions and legal constraints. That distinction carries less weight if AI governance is determined primarily through executive ultimatums issued behind closed doors.</p><p>There is also a strategic dimension. If companies conclude that participation in federal markets requires surrendering all deployment conditions, some may exit those markets. Others may respond by weakening or removing model safeguards to remain eligible for government contracts. Neither outcome strengthens <a href="https://www.reuters.com/business/retail-consumer/big-tech-group-tells-pentagons-hegseth-they-are-concerned-about-declaring-2026-03-04/" rel="noopener noreferrer" target="_blank">U.S. technological leadership</a>.</p><p>The DOD is correct that it cannot allow potential “ideological constraints” to undermine lawful military operations. But there is a difference between rejecting arbitrary restrictions and rejecting any role for corporate risk management in shaping deployment conditions. In high-risk domains—from aerospace to cybersecurity—contractors routinely impose safety standards, testing requirements, and operational limitations as part of responsible commercialization. AI should not be treated as uniquely exempt from that practice.</p><p>Moreover, built-in safeguards need not be seen as obstacles to military effectiveness. In many high-risk sectors, layered oversight is standard practice: internal controls, technical fail-safes, auditing mechanisms, and legal review operate together. Technical constraints can serve as an additional backstop, reducing the risk of misuse, error, or unintended escalation.</p><p><strong>Congress is AWOL</strong></p><p>The DOD should retain ultimate authority over lawful use. But it need not reject the possibility that certain guardrails embedded at the design level could complement its own oversight structures rather than undermine them. In some contexts, redundancy in safety systems strengthens, not weakens, operational integrity.</p><p>At the same time, a company’s unilateral ethical commitments are no substitute for public policy. When technologies carry national security implications, private governance has inherent limits. Ultimately, decisions about surveillance authorities, autonomous weapons, and rules of engagement belong in democratic institutions.</p><p>This episode illustrates a pivotal moment in AI governance. AI systems at the frontier of technology are now powerful enough to influence intelligence analysis, logistics, cyber operations and potentially battlefield decision-making. That makes them too consequential to be governed solely by corporate policy—and too consequential to be governed solely by executive discretion.</p><p>The solution is not to empower one side over the other. It is to strengthen the institutions that mediate between them.</p><p>Congress should clarify statutory boundaries for military AI use and investigate whether sufficient oversight exists. The DOD should articulate detailed doctrine for human control, auditing and accountability. Civil society and industry should participate in structured consultation processes rather than episodic standoffs and procurement policy should reflect those publicly established standards.</p><p>If AI guardrails can be removed through contract pressure, they will be treated as negotiable. However, if they are grounded in law, they can become stable expectations.</p><p>Democratic constraints on military AI belong in statute and doctrine—not in private contract negotiations.</p><p><em>This article is adapted by the author with permission from </em><a href="https://www.techpolicy.press/" rel="noopener noreferrer" target="_blank"><em>Tech Policy Press</em></a><em>. Read the </em><a href="https://www.techpolicy.press/why-congress-should-step-into-the-anthropicpentagon-dispute/" rel="noopener noreferrer" target="_blank"><em>original article</em></a><em>.</em></p>]]></description><pubDate>Sun, 08 Mar 2026 10:00:03 +0000</pubDate><guid>https://spectrum.ieee.org/military-ai-governance</guid><category>Ai</category><category>Military-ai</category><category>Anthropic</category><dc:creator>Daniel Castro</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-white-man-in-his-40s-speaking-into-a-microphone-he-is-wearing-glasses-a-suit-jacket-and-tie.jpg?id=65162768&amp;width=980"></media:content></item><item><title>Entomologists Use a Particle Accelerator to Image Ants at Scale</title><link>https://spectrum.ieee.org/3d-scanning-particle-accelerator-antscan</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/four-grey-3d-models-of-ants-shown-up-close-in-high-detail-two-larger-ants-tower-above-two-smaller-ones-in-the-front-the-larges.jpg?id=65150255&width=1200&height=800&coordinates=62%2C0%2C63%2C0"/><br/><br/><p>Move over, Pixar. The ants that animators once morphed into googly-eyed caricatures in films such as <em>A Bug’s Life</em> and <em>Antz</em> just received a meticulously precise anatomical reboot.</p><p><a href="https://doi.org/10.1038/s41592-026-03005-0" rel="noopener noreferrer" target="_blank">Writing today in <em>Nature Methods</em></a>, an international team of entomologists, accelerator physicists, computer scientists, and biological-imaging specialists describe a new 3D atlas of ant morphology.</p><p>Dubbed Antscan, the platform features micrometer-resolution reconstructions that lay bare not only the <a href="https://spectrum.ieee.org/festo-bionic-ants-and-butterflies" target="_self">insects’ armored exoskeletons</a> but also their muscles, nerves, digestive tracts, and needlelike stingers poised at the ready.</p><p>Those high-resolution images—spanning 792 species across 212 genera and covering the bulk of described ant diversity—are now available free of charge through an <a href="http://www.antscan.info" rel="noopener noreferrer" target="_blank">interactive online portal</a>, where anyone can rotate, zoom, and virtually “dissect” the insects from a laptop.</p><p>“Antscan is exciting!” says <a href="https://experts.mcmaster.ca/people/curric7" rel="noopener noreferrer" target="_blank">Cameron Currie</a>, an evolutionary biologist at McMaster University in Hamilton, Ont., Canada, who was not involved in the research. “It provides an outstanding resource for comparative work across ants.”</p><h2>Digital Access to Natural History Collections</h2><p>It also provides broader access to natural history collections.</p><p>No longer must these vast archives of preserved life be confined to drawers and jars in museums scattered around the world, available only to specialists able to visit in person. All these specimens can now be explored digitally by anyone with an internet connection, adding fresh scientific value to museum holdings.</p><p>“The more people that access and work with the stuff in our museums, whether it’s physically or digitally, the greater value they add,” says <a href="https://www.floridamuseum.ufl.edu/blackburn-lab/personnel/principal-investigator/" rel="noopener noreferrer" target="_blank">David Blackburn</a>, the curator of herpetology at the Florida Museum of Natural History who, like Currie, was not involved in the research.</p><p>Some of those people may be professional myrmecologists (scientists who specialize in the study of ants) and fourmiculture (ant-farming) enthusiasts. But others may be schoolteachers, video-game designers, tattoo artists, or curious members of the public.</p><p>“It is an extremely rich dataset that can be used for a number of different applications in science, but  also for the arts and outreach and education.” says <a href="https://www.oist.jp/image/julian-katzke" rel="noopener noreferrer" target="_blank"><span>Julian Katzke</span></a>, an entomologist at the National Museum of Natural History in Washington, D.C.</p><p>Card-carrying members of <em>IEEE</em> should find plenty to explore in Antscan as well, says <a href="https://entomology.umd.edu/people/evan-economo" target="_blank">Evan Economo</a>, a biodiversity scientist at the University of Maryland in College Park who, along with Katzke, co-led the project. <span>With the dataset now publicly available and standardized at scale, “I would really like to see these big libraries of organismal form one day be useful for people in robotics and engineering, so they can mine these data for new kinds of biomechanical designs,” he says.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Various 3D renderings of an ant soldier. First, the outward appearance. Followed by cross sectional slices of its body. One shows the internal structures of the ant, with space predominantly occupied by muscles. Another shows the same view, but with muscles removed, which highlights the digestive tract and nervous system. Lastly, zoomed-in renderings inside the ant's brain, gut and sting apparatus are shown with labels." class="rm-shortcode" data-rm-shortcode-id="672fbae791e49ff86839c1593eccc48d" data-rm-shortcode-name="rebelmouse-image" id="33b51" loading="lazy" src="https://spectrum.ieee.org/media-library/various-3d-renderings-of-an-ant-soldier-first-the-outward-appearance-followed-by-cross-sectional-slices-of-its-body-one-show.jpg?id=65150295&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">These renderings reveal different structures within the body of an army ant (<i>Eciton hamatum</i>) subsoldier, based on Antscan data.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit..."><a href="https://doi.org/10.1038/s41592-026-03005-0" target="_blank">Katzke et al.</a></small></p><h2>Advancements in Ant Imaging Technology</h2><p>Researchers have been digitizing natural history collections for years: photographing drawers of pinned specimens, building surface-level models from overlapping image stacks, and using computed tomography (CT) to scan select species one at a time. But those efforts are typically slow, piecemeal, and often limited to external features.</p><p>To capture entire organisms, inside and out, Economo and his team—then based at the Okinawa Institute of Science and Technology in Japan and including former lab members Katzke and <a href="https://www.museumfuernaturkunde.berlin/en/research/research/dynamics-nature/center-integrative-biodiversity-discovery" target="_blank">Francisco Hita Garcia</a> (now at the Museum für Naturkunde in Berlin)—built an automated imaging pipeline that effectively turned a particle accelerator into an anatomical assembly line.</p><p>They scoured museum collections around the world for ant specimens—workers, queens, and males alike—and sent some 2,200 preserved samples through a pair of micro-CT beamlines at the Karlsruhe Institute of Technology’s synchrotron <a href="https://www.ibpt.kit.edu/KIT_Light_Source.php" target="_blank">light source facility</a> in Germany.<strong></strong></p><p>There, biological imaging specialist <a href="https://www.ips.kit.edu/2890_5177.php" target="_blank">Thomas van de Kamp</a> oversaw the operation, as intense X-ray beams swept through each specimen and high-speed detectors recorded thousands of projection images from multiple angles. Robotic handlers moved vials of alcohol-preserved ants into position, one after another, all in a matter of days.</p><p>Software then reconstructed 200-plus terabytes of data generated into 3D volumes, with neural networks helping to automate the identification and analysis of anatomical structures.</p><p>Similar large-scale digitization efforts—such as the <a href="https://www.floridamuseum.ufl.edu/overt/" target="_blank">openVertebrate Project</a>, led by the Florida Museum of Natural History’s Blackburn, which involved <a href="https://academic.oup.com/bioscience/article/74/3/169/7615104" target="_blank">scanning thousands</a> of birds, fish, mammals, reptiles, and amphibians—have begun transforming how biologists study anatomy. But applying conventional micro-CT at comparable scale to insects, which are smaller and harder to scan at useful resolutions, required a leap in speed and throughput.</p><p>That’s where the synchrotron came in. By harnessing a particle accelerator to generate extraordinarily bright, coherent X-rays, the team was able to capture high-resolution internal anatomy in seconds, without the lengthy staining or other preprocessing steps often required for soft-tissue contrast in standard lab scanners.</p><p>“It is an impressive piece of work,” says <a href="https://www.nms.ac.uk/profile/dr-vladimir-blagoderov" target="_blank">Vladimir Blagoderov</a>, principal curator of invertebrates at the National Museums Scotland in Edinburgh, who was not involved in the research. “This project adds an industrial dimension to CT scanning by combining robotics, standardized sampling, automated image-processing pipelines, and machine learning.”</p><p>The sheer taxonomic breadth of the Antscan dataset now makes it possible to spot patterns across the entire ant family tree, as Economo and his colleagues have already demonstrated.</p><p>In a separate paper published last December, for example, the researchers drew on the newly generated scans to measure how much ants invest in their outer protective casing. Reporting in <em>Science Advances,</em> they showed that species with lighter, less costly cuticles <a href="https://www.science.org/doi/10.1126/sciadv.adx8068" target="_blank"><span><span>tend to form larger colonies and diversify more rapidly</span></span></a> over evolutionary time.</p><p>In their latest study, the Antscan team  turned to a different evolutionary question: The distribution of a biomineral “armor” layer <a href="https://www.nature.com/articles/s41467-020-19566-3" target="_blank">first described</a> by Currie and his colleagues in 2020 in a Central American leaf-cutter ant. A quick sweep through the Antscan database revealed that this coating—which absorbs X-rays and is visible as a bright sheath over the cuticle—is not an oddity confined to one species.</p><p>Instead, it is common among fungus-farming ants, the evolutionary lineage from which leaf-cutting ants arose roughly 20 million years ago, but largely absent in most other branches of the ant tree. (Currie’s team independently confirmed the pattern using X-ray diffraction, a technique that can precisely reveal a material’s mineral composition, as the group <a href="https://www.biorxiv.org/content/10.64898/2026.02.07.704540v1" target="_blank">reported last month in a preprint</a> posted to <em>bioRxiv</em>.)</p><p>Those are only early demonstrations of what the database can do, though. And with AI tools increasingly capable of parsing enormous, information-rich data troves, the real analytical power of Antscan may still lie ahead, says <a href="https://agsci.colostate.edu/agbio/gillette-museum/museum-staff/" target="_blank">Marek Borowiec</a>, director of the C.P. Gillette Museum of Arthropod Diversity at Colorado State University, who has chronicled <a href="https://besjournals.onlinelibrary.wiley.com/doi/full/10.1111/2041-210X.13901" target="_blank">the rise of <span>deep learning tools</span></a> in ecology and evolution.</p><p>“The full advantage of this dataset will be realized when these methods are deployed,” he says.</p><h2>Transforming Morphology with Antscan</h2><p>The ambitions behind Antscan extend well beyond ant biology. Economo and his colleagues see it as a blueprint for digitizing, standardizing, and scaling anatomy itself.<br/></p><p>Just as <a href="https://spectrum.ieee.org/whole-genome-sequencing" target="_self">large-scale sequencing projects</a> and genomic databases transformed the study of DNA over the past two decades, they hope Antscan will catalyze a comparable shift for morphology. <span>“This is kind of like having a genome for shape,” Economo says.</span></p><p>Museum collections house millions of alcohol-preserved insects and other small invertebrates—beetles, flies, wasps, spiders, crustaceans—many of them representing rare or extinct populations. Following the Antscan playbook, each could be converted into a high-resolution library of “<a data-linked-post="2655774779" href="https://spectrum.ieee.org/climate-models" target="_blank">digital twins.</a>“</p><p>In each case, synchrotron micro-CT would offer a rapid way to peer inside fragile specimens without cutting them open, capturing both hard exoskeleton and soft tissue in exquisite detail across vast swaths of biological diversity.</p><p class="shortcode-media shortcode-media-youtube"> <span class="rm-shortcode" data-rm-shortcode-id="a20837327321eee6ad3fab098e4da2e3" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/neYh_KITjGE?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span> <small class="image-media media-photo-credit" placeholder="Add Photo Credit..."><a href="https://www.youtube.com/watch?v=neYh_KITjGE" target="_blank">Antscan/YouTube</a></small></p><p><span>Beam time at major synchrotron facilities is scarce and fiercely competitive, a practical bottleneck for any effort to digitize biodiversity at scale, notes National Museums Scotland’s Blagoderov. What’s more, “even once the scans exist, the downstream burden is nontrivial: M</span><span>oving, storing, and processing hundreds of terabytes of data can become a bottleneck in its own right,” he says.</span></p><p>But if access can be secured and the computational infrastructure scaled to match, such efforts could transform natural history museums from static repositories into dynamic digital biomes.</p><p>That transformation may prove especially important at a time of accelerating species loss on Earth. By capturing organisms in extraordinary detail, resources like Antscan create a permanent, high-resolution record of life’s architecture—an anatomical time capsule that can be queried and revisited long after fragile specimens degrade or wild populations vanish.</p><p>And should Pixar ever greenlight <em>A Bug’s Life 2 </em>(suggested title: <em>Even Buggier</em>),<em> </em>the studio’s character designers may not need to take much artistic license at all. Thanks to a particle accelerator and a small cadre of dedicated scientists, the reference models are already in hand—rendered not in animation software but in micrometer-perfect anatomical form.</p>]]></description><pubDate>Thu, 05 Mar 2026 10:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/3d-scanning-particle-accelerator-antscan</guid><category>Machine-learning</category><category>Insects</category><category>Particle-accelerator</category><category>Computed-tomography</category><dc:creator>Elie Dolgin</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/four-grey-3d-models-of-ants-shown-up-close-in-high-detail-two-larger-ants-tower-above-two-smaller-ones-in-the-front-the-larges.jpg?id=65150255&amp;width=980"></media:content></item><item><title>Watershed Moment for AI–Human Collaboration in Math</title><link>https://spectrum.ieee.org/ai-proof-verification</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/four-by-four-grid-of-circles-with-varying-color-gradient-patterns.jpg?id=65103143&width=2000&height=1500&coordinates=0%2C0%2C0%2C0"/><br/><br/><p><span>When Ukrainian mathematician </span><a href="https://people.epfl.ch/maryna.viazovska?lang=en" target="_blank">Maryna Viazovska</a><span> received a </span><a href="https://www.mathunion.org/imu-awards/fields-medal/fields-medals-2022" target="_blank">Fields Medal</a><span>—widely regarded as the Nobel Prize for mathematics—in July 2022,</span><span> it was big news. Not only was she the second woman to accept the honor in the award’s 86-year history, but she collected the medal just months after her country had been invaded by Russia. Nearly four years later, Viazovska is making waves again. <a href="https://www.math.inc/sphere-packing" target="_blank">Today</a>, in </span><span>a collaboration between humans and AI, Viazovska’s proofs have been formally verified, signaling rapid progress in AI’s abilities to <a href="https://spectrum.ieee.org/ai-math-benchmarks" target="_blank">assist</a> with mathemat</span><span>ical research. </span></p><p><span>“These new results seem very, very impressive, and definitely signal some rapid progress in this direction,” says AI-reasoning expert and Princeton University postdoc <a href="https://ai.princeton.edu/news/2025/ai-lab-welcomes-associate-research-scholars" target="_blank">Liam Fowl</a>, who was not involved in the work.</span></p><p>In her Fields Medal–winning research, Viazovska had tackled two versions of the sphere-packing problem, which asks: How densely can identical circles, spheres, et cetera, be packed in <em>n</em>-dimensional space? In two dimensions, the honeycomb is the best solution. In three dimensions, spheres stacked in a pyramid are optimal. But after that, it becomes exceedingly difficult to find the best solution, and to prove that it is in fact the best. </p><p>In 2016, Viazovska solved the problem in two cases. By using powerful mathematical functions known as (quasi-)modular forms, she proved that a symmetric arrangement known as E<sub>8</sub> is the <a href="https://annals.math.princeton.edu/articles/keyword/sphere-packing" target="_blank">best 8-dimensional packing</a>, and soon after proved with collaborators that another sphere packing called the <a href="https://annals.math.princeton.edu/2017/185-3/p08" target="_blank">Leech lattice is best in 24 dimensions</a>. Though seemingly abstract, this result has potential to help solve everyday problems related to dense sphere packing, including <a data-linked-post="2650280110" href="https://spectrum.ieee.org/novel-error-correction-code-opens-a-new-approach-to-universal-quantum-computing" target="_blank">error-correcting codes</a> used by smartphones and space probes.</p><p>The proofs were verified by the mathematical community and deemed correct, leading to the Fields Medal recognition. But formal verification—the ability of a proof to be verified by a computer—is another beast altogether. Since 2022, much <a href="https://cacm.acm.org/research/formal-reasoning-meets-llms-toward-ai-for-mathematics-and-verification/" target="_blank">progress</a> has been made in AI-assisted formal proof verification. </p><h2>Serendipity leads to formalization project</h2><p>A few years later, a chance meeting in Lausanne, Switzerland, between third-year undergraduate <a href="https://thefundamentaltheor3m.github.io/" target="_blank">Sidharth Hariharan</a> and Viazovska would reignite her interest in sphere-packing proofs. Though still very early in his career, Hariharan was already becoming adept at formalizing proofs.</p><p>“Formal verification of a proof is like a rubber stamp,” Fowl says. “It’s a kind of bona fide certification that you know your statements of reasoning are correct.”</p><p>Hariharan told Viazovska how he had been using the process of formalizing proofs to learn and really understand mathematical concepts. In response, Viazovska expressed an interest in formalizing her proofs, largely out of curiosity. From this, in March 2024 the <a href="https://thefundamentaltheor3m.github.io/Sphere-Packing-Lean/" target="_blank">Formalising Sphere Packing in Lean</a> project was born. <span>Lean is a popular programming language and “proof assistant” that allows mathematicians to write proofs that are then verified for absolute correctness by a computer.</span></p><p>A collaboration formed to write a human-readable “blueprint” that could be used to map the 8-dimensional proof’s various constituents and figure out which of them had and had not been formalized and/or proven, and then prove and formalize those missing elements in Lean. </p><p><span>“We had been building the project’s repository for about 15 months when we enabled public access in June 2025,” recalls Hariharan, now a first-year Ph.D. student at Carnegie Mellon University. “Then, in late October we heard from Math, Inc. for the first time.”</span></p><h2>The AI speedup</h2><p><a href="https://www.math.inc/" target="_blank">Math, Inc.</a> is a startup developing Gauss, an AI specifically designed to automatically formalize proofs. “It’s a particular kind of language model called a reasoning agent that’s meant to interleave both traditional natural-language reasoning and fully formalized reasoning,” explains <a href="https://jesse-michael-han.github.io/" target="_blank">Jesse Han</a>, Math, Inc. CEO and cofounder. “So it’s able to conduct literature searches, call up tools, and use a computer to write down Lean code, take notes, spin up verification tooling, run the Lean compiler, et cetera.”</p><p>Math, Inc. first hit the headlines when it announced that Gauss had completed a <a href="https://mathstodon.xyz/@tao/111847680248482955" target="_blank">Lean formalization of the strong <span>prime number theorem</span> (PNT)</a> in three weeks last summer, a task that Fields Medalist <a href="https://terrytao.wordpress.com/" target="_blank">Terence Tao</a> and <a href="https://sites.math.rutgers.edu/~alexk/" target="_blank">Alex Kontorovich</a> had been working on. Similarly, Math, Inc. contacted Hariharan and colleagues to say that Gauss had proven several facts related to their sphere-packing project.</p><p>“They told us that they had finished 30 “sorrys,” which meant that they proved 30 intermediate facts that we wanted proved,” explains Hariharan. A proportion of these sorrys were shared with the project team and merged with their own work. “One of them helped us identify a typo in our project, which we then fixed,” adds Hariharan. “So it was a pretty fruitful collaboration.”</p><h2>From 8 to 24 dimensions</h2><p>But then, radio silence followed. Math, Inc. appeared to lose interest. However, while Hariharan and colleagues continued their labor of love, Math, Inc. was building a new and improved version of Gauss. “We made a research breakthrough sometime mid-January that produced a much stronger version of Gauss,” says Han. “This new version reproduced our three-week PNT result in two to three days.”</p><p>Days later, the new Gauss was steered back to the sphere-packing formalization. Working from the invaluable preexisting blueprint and work that Hariharan and collaborators had shared, Gauss not only autoformalized the 8-dimensional case, but also found and fixed a typo in the published paper, all in the space of five days.</p><p>“When they reached out to us in late January saying that they finished it, to put it very mildly, we were very surprised,” says Hariharan. “But at the end of the day, this is technology that we’re very excited about, because it has the capability to do great things and to assist mathematicians in remarkable ways.”</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="A laptop with sphere packing code in the foreground, with an autumn sunset at Carnegie Mellon in the background. " class="rm-shortcode" data-rm-shortcode-id="1dd0742602809b330ce11552ae9d6d3f" data-rm-shortcode-name="rebelmouse-image" id="898fd" loading="lazy" src="https://spectrum.ieee.org/media-library/a-laptop-with-sphere-packing-code-in-the-foreground-with-an-autumn-sunset-at-carnegie-mellon-in-the-background.jpg?id=65106120&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Hariharan was working on sphere-packing proof verification as the sun was setting behind Carnegie Mellon’s Hamerschlag Hall.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Sidharth Hariharan</small></p><p>The 8-dimensional sphere-packing proof formalization alone, <a href="https://leanprover.zulipchat.com/#narrow/channel/113486-announce/topic/Sphere.20Packing.20Milestone/with/575354368" target="_blank">announced on February 23</a>, represents a watershed moment for autoformalization and AI–human collaboration. But <a target="_blank"></a><a href="https://math.inc/sphere-packing" target="_blank">today, Math, Inc. revealed</a><span> </span>an even more impressive accomplishment: Gauss has autoformalized Viazovska’s 24-dimensional sphere-packing proof—all 200,000+ lines of code of it—in just two weeks. </p><p>There are commonalities between the 8- and 24-dimensional cases in terms of the foundational theory and overall architecture of the proof, meaning some of the code from the 8-dimensional case could be refactored and reused. However, Gauss had no preexisting blueprint to work from this time. “And it was actually significantly more involved than the 8-dimensional case, because there was a lot of missing background material that had to be brought on line surrounding many of the properties of the Leech lattice, in particular its uniqueness,” explains Han.</p><p>Though the 24-dimensional case was an automated effort, both Han and Hariharan acknowledge the many contributions from humans that laid the foundations for this achievement, regarding it as a collaborative endeavor overall between humans and AI.</p><p>But for Han, it represents even more: the beginning of a revolutionary transformation in mathematics, where extremely large-scale formalizations are commonplace. “A programmer used to be someone who punched holes into cards, but then the act of programming became separated from whatever material substrate was used for recording programs,” he concludes. “I think the end result of technology like this will be to free mathematicians to do what they do best, which is to dream of new mathematical worlds.”</p>]]></description><pubDate>Mon, 02 Mar 2026 18:00:03 +0000</pubDate><guid>https://spectrum.ieee.org/ai-proof-verification</guid><category>Mathematics</category><category>Ai-reasoning</category><category>Large-language-models</category><category>Ai</category><dc:creator>Benjamin Skuse</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/four-by-four-grid-of-circles-with-varying-color-gradient-patterns.jpg?id=65103143&amp;width=980"></media:content></item></channel></rss>