<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:media="http://search.yahoo.com/mrss/"><channel><title>IEEE Spectrum</title><link>https://spectrum.ieee.org/</link><description>IEEE Spectrum</description><atom:link href="https://spectrum.ieee.org/feeds/topic/computing.rss" rel="self"></atom:link><language>en-us</language><lastBuildDate>Thu, 09 Apr 2026 14:41:34 -0000</lastBuildDate><image><url>https://spectrum.ieee.org/media-library/eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpbWFnZSI6Imh0dHBzOi8vYXNzZXRzLnJibC5tcy8yNjg4NDUyMC9vcmlnaW4ucG5nIiwiZXhwaXJlc19hdCI6MTgyNjE0MzQzOX0.N7fHdky-KEYicEarB5Y-YGrry7baoW61oxUszI23GV4/image.png?width=210</url><link>https://spectrum.ieee.org/</link><title>IEEE Spectrum</title></image><item><title>Chip Can Project Video the Size of a Grain of Sand</title><link>https://spectrum.ieee.org/mems-photonics</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/an-array-of-tiny-metallic-cantilevers-curving-away-from-the-surface-of-a-photonic-chip.jpg?id=65493217&width=1200&height=800&coordinates=156%2C0%2C156%2C0"/><br/><br/><p><span>By many estimates, quantum computers will need <a href="https://spectrum.ieee.org/neutral-atom-quantum-computing" target="_blank">millions of qubits </a>to realize their potential applications in cybersecurity, drug development, and other industries. The problem is, anyone who has wanted to simultaneously control millions of a certain kind of qubits has run into the problem of trying to control millions of laser beams. </span> </p><p><span>That’s exactly the challenge that was faced by scientists working on the <a href="https://www.mitre.org/resources/quantum-moonshot" target="_blank">MITRE Quantum Moonshot project</a>, which brought together scientists from MITRE, MIT, the University of Colorado at Boulder, and Sandia National Laboratories. The solution they developed came in the form of an image projection technology that they realized could also be the fix for a host of other challenges in augmented reality, biomedical imaging, and elsewhere. The device is a one-square-millimeter photonic chip capable of projecting the Mona Lisa onto an area smaller than the size of two human egg <a href="https://spectrum.ieee.org/embryo-electrode-array" target="_blank">cells</a>. </span> </p><p><span>“When we started, we certainly never would have anticipated that we would be making a technology that might revolutionize imaging,” says Matt Eichenfield, one of the leaders of the Quantum Moonshot project, a collaborative research effort focused on developing a scalable diamond-based quantum computer, and a professor of quantum engineering at the University of Colorado at Boulder. Each second, their chip is capable of projecting 68.6 million individual spots of light—called scannable pixels to differentiate them from physical pixels. That’s more than fifty times the capability of previous technology, such as <a href="https://spectrum.ieee.org/mems-lidar" target="_blank">micro-electromechanical systems (MEMS) micromirror arrays</a>.</span></p><p> <span>“We have now made a scannable pixel that is at the absolute limit of what diffraction allows,” says <a href="https://www.linkedin.com/in/y-henry-wen-2b41979/" target="_blank">Henry Wen</a>, a visiting researcher at MIT and a photonics engineer at <a href="https://www.quera.com/" target="_blank">QuEra Computing</a>.</span></p><p>The chip’s distinguishing feature is an array of tiny micro-scale cantilevers, which curve away from the plane of the chip in response to voltage and act as miniature “ski-jumps” for light. Light is channeled along the length of each cantilever via a waveguide, and exits at its tip. The cantilevers contain a thin layer of aluminum nitride, a piezoelectric which expands or contracts under voltage, thus moving the micromachine up and down and enabling the array to scan beams of light over a two-dimensional area.</p><p>Despite the magnitude of the team’s achievement, Eichenfield says that the process of engineering the cantilevers was “pretty smooth.” Each cantilever is composed of a stack of several submicrometer layers of material and curls approximately 90 degrees out of the plane at rest. To achieve such a high curvature, the team took advantage of differences in the contraction and expansion of individual layers caused by physical stresses in the material resulting from the fabrication process. The materials are first deposited flat onto the chip. Then, a layer in the chip below the cantilever is removed, allowing the material stresses to take effect, releasing the cantilever from the chip and allowing it to curl out. The top layer of each cantilever also features a series of silicon dioxide bars running perpendicular to the waveguide, which keep the cantilever from curling along its width while also improving its length-wise curvature.</p><p class="shortcode-media shortcode-media-youtube"> <span class="rm-shortcode" data-rm-shortcode-id="5525c992b93704c6dfdada2cd2c1d9c2" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/A4-ZqQTZauw?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span> <small class="image-media media-caption" placeholder="Add Photo Caption...">A micro-cantilever wiggles and waggles to project light in the right place.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Matt Saha, Y. Henry Wen, et al.</small></p><p>What was more of a challenge than engineering the chip itself was figuring out the details of actually making the chip project images and videos. Working out the process of synchronizing and timing the cantilevers’ motion and light beams to generate the right colors at the right time was a substantial effort, according to <a href="https://www.linkedin.com/in/agreenspon/" target="_blank">Andy Greenspon</a>, a researcher at MITRE who also worked on the project. Now, the team has successfully projected a variety of videos from a single cantilever, including clips from the movie <em><em><a href="https://www.youtube.com/watch?v=GPG3zSgm_Qo&list=PLnvfBuirq7alZgA0yGBnNObE5CeJTpUW4" target="_blank">A Charlie Brown Christmas</a></em></em>. </p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="A warped projection of the Mona Lisa." class="rm-shortcode" data-rm-shortcode-id="a4e5294e1a010872e545dbc18fb0e208" data-rm-shortcode-name="rebelmouse-image" id="a1039" loading="lazy" src="https://spectrum.ieee.org/media-library/a-warped-projection-of-the-mona-lisa.jpg?id=65493253&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">The chip projected a roughly 125-micrometer image of the Mona Lisa.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit..."><a href="https://www.nature.com/articles/s41586-025-10038-6" target="_blank">Matt Saha, Y. Henry Wen, et al.</a></small></p><p>Because the chip can project so many more spots in any given time interval than any previous beam scanners, it could also be used to control many more qubits in quantum computers. The Quantum Moonshot program’s mission is to build a quantum computer that can be scaled to millions of qubits. So clearly, it needs a scalable way of controlling each one, explains Wen. Instead of using one laser per qubit, the team realized that not every qubit needed to be controlled at every given moment. The chip’s ability to move light beams over a two-dimensional area, would allow them to control all of the qubits with many fewer lasers. </p><p>Another process that Wen thinks the chip could improve is scanning objects for <a href="https://spectrum.ieee.org/3d-printed-linear-motor" target="_blank">3D printing</a>. Today, that typically involves using a single laser to scan over the entire surface of an object. The new chip, however, could potentially employ thousands of laser beams. “I think now you can take a process that would have taken hours and maybe bring it down to minutes,” says Wen. </p><p>Wen is also excited to explore the potential of different cantilever shapes. By changing the orientations of the bars perpendicular to the waveguide, the team has been able to make the cantilevers curl into helixes. Wen says that such unusual shapes could be useful in making a <a href="https://spectrum.ieee.org/neurobot-living-robot-nervous-system" target="_blank">lab-on-a-chip for cell biology</a> or <a href="https://spectrum.ieee.org/lab-on-a-chip-grippers" target="_blank">drug development</a>. “A lot of this stuff is imaging, scanning a laser across something, either to image it or to stimulate some response. And so we could have one of these ski jumps curl not just up, but actually curl back around, and then move around and scan over a sample,” Wen explains. “If you can imagine a structure that will be useful for you, we should try it.”</p>]]></description><pubDate>Thu, 09 Apr 2026 13:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/mems-photonics</guid><category>Microarray</category><category>Digital-micromirror-device</category><category>Mems</category><category>Quantum-computers</category><category>Nitrogen-vacancy-defects-diamond</category><dc:creator>Velvet Wu</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/an-array-of-tiny-metallic-cantilevers-curving-away-from-the-surface-of-a-photonic-chip.jpg?id=65493217&amp;width=980"></media:content></item><item><title>AI Models Trained on Physics Are Changing Engineering</title><link>https://spectrum.ieee.org/large-physics-models-design-engineering</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/diagram-of-airflow-over-a-moving-sedan.jpg?id=65494121&width=1200&height=800&coordinates=62%2C0%2C63%2C0"/><br/><br/><p>Large language models have already <a href="https://spectrum.ieee.org/best-ai-coding-tools" target="_self">transformed</a> software engineering, for better or worse. Now, so-called large physics models are also starting to transform design engineering. These tools are beginning to replace—or at least, amend—the role of full-fledged physics simulation in the automotive and aerospace industries, semiconductor engineering, and more.</p><p>Before the advent of computer simulation, a car manufacturer, for example, would create prototypes to test their designs, says <a href="https://www.linkedin.com/in/thomas-von-tschammer/" rel="noopener noreferrer" target="_blank">Thomas Von Tschammer</a>, managing director at physics-based AI company <a href="https://www.neuralconcept.com/" rel="noopener noreferrer" target="_blank">Neural Concept</a>. “For the past 40 years, we reduced a lot of the need for prototypes by using numerical simulations for aerodynamics, for crash testing, and so on.” Now, Tschammer explains, AI is drastically reducing the need for simulation, the same way simulation reduced the need for physical prototypes.</p><p>Growing adoptions of this type of AI was a topic of interest at <a href="https://www.nvidia.com/gtc/" rel="noopener noreferrer" target="_blank">Nvidia GTC</a> in March. <a href="https://www.linkedin.com/in/chris-johnston-/" rel="noopener noreferrer" target="_blank">Chris Johnston</a>, senior technical specialist at Jaguar Land Rover, <a href="https://www.nvidia.com/en-us/on-demand/session/gtc26-s81736/?playlistId=gtc26-industrial-engineering" rel="noopener noreferrer" target="_blank">presented</a> how his company is using Neural Concept’s technology. <a href="https://www.physicsx.ai/" rel="noopener noreferrer" target="_blank">PhysicsX</a>, another physics-based AI company, <a href="https://www.physicsx.ai/newsroom/physicsx-announces-advancement-to-open-standards-for-physics-ai-powered-by-nvidia" rel="noopener noreferrer" target="_blank">announced</a> a collaboration with Nvidia to advance open standards for such models, also at GTC.</p><h2>The AI design engineering workflow</h2><p>Over the past six months, <a href="https://www.gm.com/" rel="noopener noreferrer" target="_blank">General Motors</a> (GM) has introduced large physics models into their car design process to speed up the workflow. </p><p>Previously, a creative design engineer would develop a 3D model of a new car concept. This model would be sent to aerodynamics specialists, who would run physics simulations to determine the coefficient of drag of the proposed car—an important metric for energy efficiency of the vehicle. This simulation phase would take about two weeks, and the aerodynamics engineer would then report the drag coefficient back to the creative designer, possibly with suggested modifications.</p><p>Now, GM has trained an in-house large physics model on those simulation results. The AI takes in a 3D car model and outputs a coefficient of drag in a matter of minutes. “We have experts in the aerodynamics and the creative studio now who can sit together and iterate instantly to make decisions [about] our future products,” says <a href="https://www.linkedin.com/in/rdstrauss/" rel="noopener noreferrer" target="_blank">Rene Strauss</a>, director of virtual integration engineering at GM. </p><p>For GM and other companies, running inference on an AI model trained on physics simulations, instead of running the simulation itself, can bring immense time savings. “Depending on the kinds of physics [being simulated], or the resolution, it can be anywhere between 10,000 to close to a million times faster,” says <a href="https://www.linkedin.com/in/jacomo-corbo/" rel="noopener noreferrer" target="_blank">Jacomo Corbo</a>, CEO and co-founder of PhysicsX.</p><h2>How accurate are large physics models?</h2><p>But, what about accuracy? For GM’s purposes, Strauss says accuracy is not a huge concern at the design stage because finer details are ironed out later in the process. “When it really starts to matter is when we’re getting close to launching a vehicle, and the coefficient of drag is going to be used for our energy calculation, which eventually goes to the certification of our miles per gallon on the sticker.” At that stage, Strauss says, a physical model of the car will be put into a wind tunnel for an exact number.</p><p>PhysicsX’s Corbo argues that, with the right data, the AI model accuracy can supersede the accuracy of the simulation it’s trained on. The trick is to incorporate experimental measurements to fine-tune the model. If a physics simulation doesn’t agree exactly with experimental data, it is often difficult to figure out why and tweak the model until they agree. With AI, incorporating a few experimental examples into the training process is a lot more straightforward, and it’s not necessary to understand where exactly the model went wrong.</p><p>All in all, by drastically bringing down the time it takes to model the physics, large physics models enable engineers to explore a much greater range of possibilities before a final design is reached. </p><h2>Training large physics models</h2><p>There is no one-size-fits-all approach to training large physics models. Depending on the types of data available, and the physics in question, the models may use the <a href="https://spectrum.ieee.org/what-is-generative-ai" target="_self">transformer</a> architecture that underlies LLMs, a generalized version of convolutional neural networks known as <a href="https://dataroots.io/blog/a-gentle-introduction-to-geometric" rel="noopener noreferrer" target="_blank">geometric deep learning</a>, or an architecture that can solve partial differential equations called <a href="https://zongyi-li.github.io/neural-operator/" rel="noopener noreferrer" target="_blank">neural operators</a>.</p><p>Currently, most companies are training their own models on their simulation data, catering to specific use cases. In GM’s aerodynamics implementation, there are different AI models for different types of cars: think SUVs versus sedans. But PhysicsX’s Corbo says his team is working on building more “foundational” physics models that can be applied across different scenarios.</p><p>Both <a href="https://arxiv.org/pdf/2001.08361" rel="noopener noreferrer" target="_blank">LLMs</a> and <a href="https://spectrum.ieee.org/solve-robotics" target="_self">robotics</a> have benefitted from scaling laws, which describe how a system improves as the models increase in size or get trained on more data. In AI, models tend to improve quickly, in a non-linear way. Along the way, the models also become more generalizable—extending them to new settings takes less and less fine-tuning to reach the same accuracy. Corbo says his team is now starting to see the same types of scaling laws for large physics models.</p><p>“What we’re seeing here is maybe a little bit unsurprising,” Corbo says, “but it’s also pretty incredible. And it’s given us the confidence to make these models bigger, because they perform a whole lot better, and they cover broader domains, and they have these really amazing emergent properties.”</p><p>Developing open standards for the data formats used in training, as well as the model architectures, should help develop these more powerful foundational models. That’s the goal of PhysicsX’s collaboration with Nvidia, and of Nvidia’s <a href="https://developer.nvidia.com/blog/physics-ml-platform-physicsnemo-is-now-open-source/" rel="noopener noreferrer" target="_blank">physicsNeMo</a> open source platform.</p><p>“The thing that we’re collaborating on is being able to compose architectures from building blocks,” Corbo says, making it easy for those in both academia and industry to re-use and build upon existing models.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A physics-based AI software being used to generate a 3D geometry of a data center server PCB ready to be run in computational fluid dynamics." class="rm-shortcode" data-rm-shortcode-id="06165df81787c02f03f9299373dd26f0" data-rm-shortcode-name="rebelmouse-image" id="70269" loading="lazy" src="https://spectrum.ieee.org/media-library/a-physics-based-ai-software-being-used-to-generate-a-3d-geometry-of-a-data-center-server-pcb-ready-to-be-run-in-computational-fl.jpg?id=65494125&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">A type of AI called a large physics model is used by an to quickly generate heat flow in a 3D data center server design. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Neural Concept</small></p><h2>The long-term role of simulations and engineers</h2><p>While some are working on developing more powerful models, others are pushing to implement what’s already available into existing workflows, which is no easy task. “With any innovation, it’s not a straight line. There’s some steps forward and then some steps back and improvements that we find along the way. But that’s part of the joy of the innovation process and using new tools like this,” GM’s Strauss says.</p><p>This technology is still in the early stages, and it’s unclear what the final role of AI tools will be in the engineering workflow. For one, opinions vary on whether AI will replace simulations completely, or just reduce their use.</p><p>“We will never fully replace simulations,” Neural Concept’s von Tschammer says. “But the idea is to make a much smarter usage of simulation at the most major phase of developments, and you use AI to speed up the early design stages, where you need to explore a very wide set of options.”</p><p>PhysicsX’s Corbo begs to differ. “The whole idea is to take numerical simulation … out of the workflow,” he says, “and to move that to inference.”</p><p>Whatever the role of simulation will be, everyone in the field is adamant that human design engineers will continue to be in the driver’s seat, enabled to do their best work by these newfangled tools. (After all, when has AI ever threatened to replace human labor?)</p><p>“What we’re seeing is that actually, these tools are empowering the engineers to be much more efficient,” Tschammer says. “Before, these engineers would spend a lot of time on low added value tasks, whereas now these manual tasks from the past can be automated using these AI models, and the engineers can focus on taking the design decisions at the end of the day. We still need engineers more than ever.”</p>]]></description><pubDate>Thu, 09 Apr 2026 11:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/large-physics-models-design-engineering</guid><category>Physics-simulations</category><category>General-motors</category><category>Nvidia-gtc</category><category>Engineering-design</category><dc:creator>Dina Genkina</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/diagram-of-airflow-over-a-moving-sedan.jpg?id=65494121&amp;width=980"></media:content></item><item><title>Decentralized Training Can Help Solve AI’s Energy Woes</title><link>https://spectrum.ieee.org/decentralized-ai-training-2676670858</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/illustration-of-several-data-servers-interconnected-across-long-distances.jpg?id=65477795&width=1200&height=800&coordinates=156%2C0%2C156%2C0"/><br/><br/><p> <a href="https://spectrum.ieee.org/topic/artificial-intelligence/" target="_self">Artificial intelligence</a> harbors an enormous <a href="https://spectrum.ieee.org/topic/energy/" target="_self">energy</a> appetite. Such constant cravings are evident in the <a href="https://spectrum.ieee.org/ai-index-2025" target="_self">hefty carbon footprint</a> of the <a href="https://spectrum.ieee.org/tag/data-centers" target="_self">data centers</a> behind the AI boom and the steady increase over time of <a href="https://spectrum.ieee.org/tag/carbon-emissions" target="_self">carbon emissions</a> from training frontier <a href="https://spectrum.ieee.org/tag/ai-models" target="_self">AI models</a>.</p><p>No wonder big tech companies are warming up to <a href="https://spectrum.ieee.org/tag/nuclear-energy" target="_self">nuclear energy</a>, envisioning a future fueled by reliable, carbon-free sources. But while <a href="https://spectrum.ieee.org/nuclear-powered-data-center" target="_self">nuclear-powered data centers</a> might still be years away, some in the research and industry spheres are taking action right now to curb AI’s growing energy demands. They’re tackling training as one of the most energy-intensive phases in a model’s life cycle, focusing their efforts on decentralization.</p><p>Decentralization allocates model training across a network of independent nodes rather than relying on one platform or provider. It allows compute to go where the energy is—be it a dormant server sitting in a research lab or a computer in a <a href="https://spectrum.ieee.org/tag/solar-power" target="_self">solar-powered</a> home. Instead of constructing more data centers that require <a href="https://spectrum.ieee.org/tag/power-grid" target="_self">electric grids</a> to scale up their infrastructure and capacity, decentralization harnesses energy from existing sources, avoiding adding more power into the mix.</p><h2>Hardware in harmony</h2><p>Training AI models is a huge data center sport, synchronized across clusters of closely connected <a href="https://spectrum.ieee.org/tag/gpus" target="_self">GPUs</a>. But as <a href="https://spectrum.ieee.org/mlperf-trends" target="_self">hardware improvements struggle to keep up</a> with the swift rise in size of <a href="https://spectrum.ieee.org/tag/large-language-models" target="_self">large language models</a>, even massive single data centers are no longer cutting it.</p><p>Tech firms are turning to the pooled power of multiple data centers—no matter their location. <a href="https://spectrum.ieee.org/tag/nvidia" target="_self">Nvidia</a>, for instance, launched the <a href="https://developer.nvidia.com/blog/how-to-connect-distributed-data-centers-into-large-ai-factories-with-scale-across-networking/" target="_blank">Spectrum-XGS Ethernet for scale-across networking</a>, which “can deliver the performance needed for large-scale single job AI training and inference across geographically separated data centers.” Similarly, <a href="https://spectrum.ieee.org/tag/cisco" target="_self">Cisco</a> introduced its <a href="https://blogs.cisco.com/sp/the-new-benchmark-for-distributed-ai-networking" target="_blank">8223 router</a> designed to “connect geographically dispersed AI clusters.”</p><p>Other companies are harvesting idle compute in <a href="https://spectrum.ieee.org/tag/servers" target="_self">servers</a>, sparking the emergence of a <a href="https://spectrum.ieee.org/gpu-as-a-service" target="_self">GPU-as-a-Service</a> business model. Take <a href="https://akash.network/" rel="noopener noreferrer" target="_blank">Akash Network</a>, a peer-to-peer <a href="https://spectrum.ieee.org/tag/cloud-computing" target="_self">cloud computing</a> marketplace that bills itself as the “Airbnb for data centers.” Those with unused or underused GPUs in offices and smaller data centers register as providers, while those in need of computing power are considered as tenants who can choose among providers and rent their GPUs.</p><p>“If you look at [AI] training today, it’s very dependent on the latest and greatest GPUs,” says Akash cofounder and CEO <a href="https://www.linkedin.com/in/gosuri" rel="noopener noreferrer" target="_blank">Greg Osuri</a>. “The world is transitioning, fortunately, from only relying on large, high-density GPUs to now considering smaller GPUs.”</p><h2>Software in sync</h2><p>In addition to orchestrating the <a href="https://spectrum.ieee.org/tag/hardware" target="_self">hardware</a>, decentralized AI training also requires algorithmic changes on the <a href="https://spectrum.ieee.org/tag/software" target="_self">software</a> side. This is where <a href="https://cloud.google.com/discover/what-is-federated-learning" rel="noopener noreferrer" target="_blank">federated learning</a>, a form of distributed <a href="https://spectrum.ieee.org/tag/machine-learning" target="_self">machine learning</a>, comes in.</p><p>It starts with an initial version of a global AI model housed in a trusted entity such as a central server. The server distributes the model to participating organizations, which train it locally on their data and share only the model weights with the trusted entity, explains <a href="https://www.csail.mit.edu/person/lalana-kagal" rel="noopener noreferrer" target="_blank">Lalana Kagal</a>, a principal research scientist at <a href="https://www.csail.mit.edu/" rel="noopener noreferrer" target="_blank">MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL)</a> who leads the <a href="https://www.csail.mit.edu/research/decentralized-information-group-dig" rel="noopener noreferrer" target="_blank">Decentralized Information Group</a>. The trusted entity then aggregates the weights, often by averaging them, integrates them into the global model, and sends the updated model back to the participants. This collaborative training cycle repeats until the model is considered fully trained.</p><p>But there are drawbacks to distributing both data and computation. The constant back and forth exchanges of model weights, for instance, result in high communication costs. Fault tolerance is another issue.</p><p>“A big thing about AI is that every training step is not fault-tolerant,” Osuri says. “That means if one node goes down, you have to restore the whole batch again.”</p><p>To overcome these hurdles, researchers at <a href="https://deepmind.google/" rel="noopener noreferrer" target="_blank">Google DeepMind</a> developed <a href="https://arxiv.org/abs/2311.08105" rel="noopener noreferrer" target="_blank">DiLoCo</a>, a distributed low-communication optimization <a href="https://spectrum.ieee.org/tag/algorithms" target="_self">algorithm</a>. DiLoCo forms what <a href="https://spectrum.ieee.org/tag/google-deepmind" target="_self">Google DeepMind</a> research scientist <a href="https://arthurdouillard.com/" rel="noopener noreferrer" target="_blank">Arthur Douillard</a> calls “islands of compute,” where each island consists of a group of <a href="https://spectrum.ieee.org/tag/chips" target="_self">chips</a>. Every island holds a different chip type, but chips within an island must be of the same type. Islands are decoupled from each other, and synchronizing knowledge between them happens once in a while. This decoupling means islands can perform training steps independently without communicating as often, and chips can fail without having to interrupt the remaining healthy chips. However, the team’s experiments found diminishing performance after eight islands.</p><p>An improved version dubbed <a href="https://arxiv.org/abs/2501.18512" rel="noopener noreferrer" target="_blank">Streaming DiLoCo</a> further reduces the bandwidth requirement by synchronizing knowledge “in a streaming fashion across several steps and without stopping for communicating,” says Douillard. The mechanism is akin to watching a video even if it hasn’t been fully downloaded yet. “In Streaming DiLoCo, as you do computational work, the knowledge is being synchronized gradually in the background,” he adds.</p><p>AI development platform <a href="https://www.primeintellect.ai/" rel="noopener noreferrer" target="_blank">Prime Intellect</a> implemented a variant of the DiLoCo algorithm as a vital component of its 10-billion-parameter <a href="https://www.primeintellect.ai/blog/intellect-1-release" rel="noopener noreferrer" target="_blank">INTELLECT-1</a> model trained across five countries spanning three continents. Upping the ante, <a href="https://0g.ai/" rel="noopener noreferrer" target="_blank">0G Labs</a>, makers of a decentralized AI <a href="https://spectrum.ieee.org/tag/operating-system" target="_self">operating system</a>, <a href="https://0g.ai/blog/worlds-first-distributed-100b-parameter-ai" rel="noopener noreferrer" target="_blank">adapted DiLoCo to train a 107-billion-parameter foundation model</a> under a network of segregated clusters with limited bandwidth. Meanwhile, popular <a href="https://spectrum.ieee.org/tag/open-source" target="_self">open-source</a> <a href="https://spectrum.ieee.org/tag/deep-learning" target="_self">deep learning</a> framework <a href="https://pytorch.org/projects/pytorch/" rel="noopener noreferrer" target="_blank">PyTorch</a> included DiLoCo in its <a href="https://meta-pytorch.org/torchft/" rel="noopener noreferrer" target="_blank">repository of fault tolerance techniques</a>.</p><p>“A lot of engineering has been done by the community to take our DiLoCo paper and integrate it in a system learning over consumer-grade internet,” Douillard says. “I’m very excited to see my research being useful.”</p><h2>A more energy-efficient way to train AI</h2><p>With hardware and software enhancements in place, decentralized AI training is primed to help solve AI’s energy problem. This approach offers the option of training models “in a cheaper, more resource-efficient, more energy-efficient way,” says MIT CSAIL’s Kagal.</p><p>And while Douillard admits that “training methods like DiLoCo are arguably more complex, they provide an interesting tradeoff of system efficiency.” For instance, you can now use data centers across far apart locations without needing to build ultrafast bandwidth in between. Douillard adds that fault tolerance is baked in because “the blast radius of a chip failing is limited to its island of compute.”</p><p>Even better, companies can take advantage of existing underutilized processing capacity rather than continuously building new energy-hungry data centers. Betting big on such an opportunity, Akash created its <a href="https://www.youtube.com/watch?v=zAj41xSNPeI" rel="noopener noreferrer" target="_blank">Starcluster program</a>. One of the program’s aims involves tapping into solar-powered homes and employing the desktops and laptops within them to train AI models. “We want to convert your home into a fully functional data center,” Osuri says.</p><p>Osuri acknowledges that participating in Starcluster will not be trivial. Beyond solar panels and devices equipped with consumer-grade GPUs, participants would also need to invest in <a href="https://spectrum.ieee.org/tag/batteries" target="_self">batteries</a> for backup power and redundant internet to prevent downtime. The Starcluster program is figuring out ways to package all these aspects together and make it easier for homeowners, including collaborating with industry partners to subsidize battery costs.</p><p>Backend work is already underway to enable <a href="https://akash.network/roadmap/aep-60/" rel="noopener noreferrer" target="_blank">homes to participate as providers in the Akash Network</a>, and the team hopes to reach its target by 2027. The Starcluster program also envisions expanding into other solar-powered locations, such as schools and local community sites.</p><p>Decentralized AI training holds much promise to steer AI toward a more environmentally sustainable future. For Osuri, such potential lies in moving AI “to where the energy is instead of moving the energy to where AI is.”</p>]]></description><pubDate>Tue, 07 Apr 2026 14:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/decentralized-ai-training-2676670858</guid><category>Training</category><category>Ai-energy</category><category>Data-center</category><category>Large-language-models</category><dc:creator>Rina Diane Caballar</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/illustration-of-several-data-servers-interconnected-across-long-distances.jpg?id=65477795&amp;width=980"></media:content></item><item><title>ENIAC’s Architects Wove Stories Through Computing</title><link>https://spectrum.ieee.org/eniac-80th-anniversary-weaving</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/close-up-black-and-white-1940-s-image-of-a-woman-holding-a-metallic-brick-like-controller-with-large-knobs.jpg?id=65453792&width=1200&height=800&coordinates=0%2C83%2C0%2C84"/><br/><br/><p><em><em>This year marks the </em></em><a href="https://spectrum.ieee.org/eniac-80-ieee-milestone" target="_self"><em><em>80th anniversary of ENIAC</em></em></a><em><em>, the first general-purpose digital computer. The computer was built during World War II to speed up ballistics calculations, but its contributions to computing extend well beyond military applications. </em></em></p><p><em><em>Two of ENIAC’s key architects—John W. Mauchly, its co-inventor, and Kathleen “Kay” McNulty, one of the <a href="https://spectrum.ieee.org/eniac-woman-programmers" target="_blank">six original programmers</a>—married a few years after its completion and raised seven children together. Mauchly and McNulty’s grandchild Naomi Most </em></em><a href="https://youtu.be/XYEVmqGhVxo?si=fseDLKFz1W8meWR6&t=4515" rel="noopener noreferrer" target="_blank"><em><em>delivered a talk</em></em></a><em><em> as part of a celebration in honor of ENIAC’s anniversary on 15 February, which was held online and in-person at the American Helicopter Museum in West Chester, Pa. The following is adapted from that presentation.</em></em></p><p class="ieee-inbody-related">RELATED: <a href="https://spectrum.ieee.org/eniac-80-ieee-milestone" target="_blank">ENIAC, the First General-Purpose Digital Computer, Turns 80</a></p><p>There was a library at my grandparents’ farmhouse that felt like it went on forever. September light through the windows, beech leaves rustling outside on the stone porch, the sounds of cousins and aunts and uncles somewhere in the house. And in the corner of that library, an IBM personal computer.</p><p>When I spent summers there as a child, I didn’t yet know that the computer was closely tied to my family’s story.</p><p>My grandparents are known for their contributions to creating the Electronic Numerical Integrator and Computer, or ENIAC. But both were interested in more than just crunching numbers: My grandfather wanted to predict the weather. My grandmother wanted to be a good storyteller. </p><p>In Irish, the first language my grandmother Kathleen “Kay” McNulty ever spoke, a word existed to describe both of these impulses: <em><em>ríomh</em></em>.</p><p>I began to learn the Irish language myself five years ago, and I was struck by how certain words and phrases had multiple meanings. According to renowned Irish cultural historian Manchán Magan—from whom I took lessons—the word <em><em>ríomh</em></em> has at different times been used to mean to compute, but also <a href="https://www.making.ie/stories/irish-words-weaving" rel="noopener noreferrer" target="_blank">to weave, to narrate, or to compose a poem</a>. That one word that can tell the story of ENIAC, a machine with wires woven like thread that was built to compute, make predictions, and search for a signal in the noise. </p><h2>John Mauchly’s Weather-Prediction Ambitions</h2><p>Before working on ENIAC, John Mauchly <a href="https://fi.edu/en/news/case-files-john-w-mauchly-and-j-presper-eckert" rel="noopener noreferrer" target="_blank">spent years collecting rainfall data</a> across the United States. His favorite pastime was meteorology, and he wanted to find patterns in storm systems to predict the weather.</p><p>The Army, however, funded ENIAC to make simpler predictions: calculating ballistic trajectory tables. Start there, co-inventors J. Presper Eckert and Mauchly realized, and perhaps the weather would soon be computable.</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="Black and white 1960s image of two white men in suits looking at a wall of computer controls." class="rm-shortcode" data-rm-shortcode-id="7872d50df109149c936e400909defc38" data-rm-shortcode-name="rebelmouse-image" id="75108" loading="lazy" src="https://spectrum.ieee.org/media-library/black-and-white-1960s-image-of-two-white-men-in-suits-looking-at-a-wall-of-computer-controls.jpg?id=65428294&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Co-inventors John Mauchly [left] and J. Presper Eckert look at a portion of ENIAC on 25 November 1966. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Hulton Archive/Getty Images</small></p><p>Weather is a system unfolding through time, and a model of a storm is a story about how that system might unfold. There’s an old Irish saying related to this idea: <a href="https://daltai.com/is-maith-an-scealai-an-aimsir/" target="_blank"><em><em>Is maith an scéalaí an aimsir</em></em></a><em><em>.</em></em> Literally, “weather is a good storyteller.” But <em><em>aimsir</em></em> also means time. So the usual translation of this phrase into English becomes “time will tell.”</p><p>Mauchly wanted to <em><em>ríomh an aimsire</em></em>—to weave the weather into pattern, to compute the storm, to narrate the chaos. He realized that complex systems don’t reveal their full purpose at conception. They reveal it through <em><em>aimsir</em></em>—through weather, through time, through use.</p><h2>ENIAC’s First Programmers Were Weavers</h2><p>Kathleen “Kay” McNulty was born on 12 February 1921, in Creeslough, Ireland, on the night <a href="https://en.wikipedia.org/wiki/James_McNulty_(Irish_activist)" target="_blank">her father</a>—an IRA training officer—was arrested and imprisoned in Derry Gaol.</p><p>Family oral history holds that her people were weavers. She spoke only Irish until her family reached Philadelphia when she was 4 years old, entering American school the following year knowing virtually no English. She graduated in 1942 from Chestnut Hill College with a mathematics degree, was recruited to compute artillery firing tables by hand for the U.S. Army, and was then selected—along with <a href="https://spectrum.ieee.org/the-women-behind-eniac" target="_blank">five other women</a>—to program ENIAC.</p><p>They had no manual. They had only blueprints.</p><p>McNulty and her colleagues learned ENIAC and its quirks the way you learn a loom: by touch, by memory, by routing threads of electricity into patterns. They developed embodied knowledge the designers could only approximate. They could narrow a malfunction to a specific failed vacuum tube before any technician could locate it.</p><p>McNulty and Mauchly are also credited with conceiving the subroutine, the sequence of instructions that can be repeatedly recalled to perform a task, now essential in any programming. The subroutine was not in ENIAC’s blueprints, nor in the funding proposal. The concept emerged as highly determined people extended their imagination into the machine’s affordances.</p><p>The engineers designed the loom. Weavers discovered its true capabilities.</p><p>In 1950, four years after ENIAC was switched on, Mauchly’s dream was realized as it was used in the <a href="https://www.guinnessworldrecords.com/world-records/775520-first-computer-assisted-weather-forecast" target="_blank">world’s first computer-assisted weather forecast</a>. That was made possible after Klara von Neumann and Nick Metropolis reassembled and upgraded the ENIAC with a small amount of digital program memory. The programmers who transformed the math into operational code for the ENIAC were Norma Gilbarg, Ellen-Kristine Eliassen, and Margaret Smagorinsky. Their names are not as well-known as they should be.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Black and white 1940s image of three women operating a differential analyser in a basement." class="rm-shortcode" data-rm-shortcode-id="298168a77d38fd343eeb7d4bbfc219a7" data-rm-shortcode-name="rebelmouse-image" id="aacec" loading="lazy" src="https://spectrum.ieee.org/media-library/black-and-white-1940s-image-of-three-women-operating-a-differential-analyser-in-a-basement.jpg?id=65453828&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Before programming ENIAC, Kay McNulty [left] was recruited by the U.S. Army to compute artillery firing tables. Here, she and two other women, Alyse Snyder [center] and Sis Stump, operate a mechanical analog computer designed to solve differential equations in the basement of the University of Pennsylvania’s Moore School of Electrical Engineering.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">University of Pennsylvania</small></p><h2>Kay McNulty, Family Storyteller</h2><p>Kay married John Mauchly in 1948, describing him as “the greatest delight of my life. He was so intelligent and had so many ideas.... He was not only lovable, he was loving.” She spent the rest of her life ensuring he, Eckert, and the ENIAC programmers would be recognized.</p><p>When she died in 2006, I came to her funeral in shock, not fully knowing what I’d lost. As she drifted away, it was said, she had been reciting her prayers in Irish. This understanding made it quickly over to Creeslough, in County Donegal, and awaited me when I visited to honor her memory with the <a href="https://www.youtube.com/watch?v=zbkk2RJMW9g" target="_blank">dedication of a plaque</a> right there in the center of town.</p><p>In <a href="https://mathshistory.st-andrews.ac.uk/Extras/Mauchly_Antonelli_story" target="_blank">her own memoir</a>, she wrote: “If I am remembered at all, I would like to be remembered as my family storyteller.”</p><p>In Irish, the word for computer is <em><em>ríomhaire</em></em>. One who ríomhs. One who weaves, computes, and tells. My grandfather wanted to tell the story of the weather through computing. My grandmother wanted to be remembered as a storyteller. The language of her childhood already had a word that contained both of those ambitions.</p><h2>Computers as Narrative Engines</h2><p>When it was built, ENIAC looked like the back room of a textile production house. Panels. Switchboards. A room full of wires. Thread.</p><p>Thread does not tell you what it will become. We tend to think of computing as calculation—discrete and deterministic. But a model is a structured story about how something behaves.</p><p>Weather models, ballistic tables, economic forecasts, neural networks: These are all narrative engines, systems that take raw inputs and produce accounts of how the world might unfold. In complex systems, when parts are woven together through use, new structures arise that no one specified in advance.</p><p>Like ENIAC, the machines we are building now—the large models, the autonomous systems—are not merely calculators. They are looms.</p><p>Their most important properties will not be specified in advance. They will emerge through use, through the people who learn how to weave with them.</p><p>Through imagination.</p><p>Through <em><em>aimsir</em></em>.</p>]]></description><pubDate>Fri, 03 Apr 2026 13:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/eniac-80th-anniversary-weaving</guid><category>Eniac</category><category>Weather-prediction</category><category>Computer-history</category><category>Ireland</category><dc:creator>Naomi Most</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/close-up-black-and-white-1940-s-image-of-a-woman-holding-a-metallic-brick-like-controller-with-large-knobs.jpg?id=65453792&amp;width=980"></media:content></item><item><title>The AI Data Centers That Fit on a Truck</title><link>https://spectrum.ieee.org/modular-data-center</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/overhead-view-of-two-data-center-pods-each-measuring-55-feet-long-by-12-5-feet-wide.jpg?id=65417343&width=1200&height=800&coordinates=0%2C83%2C0%2C84"/><br/><br/><p>A <a data-linked-post="2676577917" href="https://spectrum.ieee.org/5gw-data-center" target="_blank">traditional</a> data center protects the expensive hardware inside it with a “shell” constructed from steel and concrete. Constructing a data center’s shell is inexpensive compared to the cost of the hardware and infrastructure inside it, but it’s not trivial. It takes time for engineers to consider potential sites, apply for permits, and coordinate with construction contractors.</p><p>That’s a problem for those looking to quickly deploy AI hardware, which has led companies like <a href="https://duosedge.ai/home" target="_blank">Duos Edge AI</a> and <a href="https://www.lgcns.com/en" target="_blank">LG CNS</a> to respond with a more modular approach. They use pre-fabricated, self-contained boxes that can be deployed in months instead of years. The boxes can operate alone or in tandem with others, providing the option to add more if required.</p><p>“I just came back from Nvidia’s GTC, and a lot of [companies] are sitting on their deployment because their data centers aren’t ready, or they can’t find the space,” said <a href="https://www.linkedin.com/in/doug-recker/" rel="noopener noreferrer" target="_blank">Doug Recker</a>, CEO of Duos Edge AI. “We see the demand there, and we can deploy faster.” </p><h2>GPUs shipped straight to you</h2><p>Duos Edge AI’s modular compute pods are 55 feet long and 12.5 feet wide. Though they look similar to a shipping container, they’re actually a bit larger and designed primarily for transportation by truck. Each compute pod contains racks of GPUs much like those used in other data centers. Duos recently <a href="https://ir.duostechnologies.com/news-events/press-releases/detail/830/duos-technologies-group-executes-definitive-agreement-with" target="_blank">entered</a> a deal with AI infrastructure company Hydra Host to deploy four pods with 576 GPUs per pod. That’s a total of 2,304 GPUs, with the option to later double the deployment to 4,608 GPUs. </p><p>Modular data centers aren’t new for Duos; the company previously deployed edge data centers for rural customers, <a href="https://spectrum.ieee.org/rural-data-centers" target="_self">such as the Amarillo, Texas school district</a>. However, the pods for the Hydra Host deployment will be upgraded to handle more intense AI workloads. They’ll contain more racks, draw more power, and use liquid cooling to keep the GPUs running efficiently. <br/><br/>Across the Pacific, Korean technology giant LG is taking a similar approach. The company’s CNS subsidiary, which provides IT infrastructure and services, <a href="https://www.koreatimes.co.kr/business/tech-science/20260305/lg-cns-unveils-container-based-ai-box-for-rapid-ai-data-center-expansion">has announced the AI Modular Data Center which</a>, like the Duos unit, contains racks of GPUs and supporting hardware in a pre-fabricated enclosure.</p><p>Also like Duos’ deployment, LG’s AI Modular Data Center contains 576 Nvidia GPUs with the option to scale up in the future. “We are currently developing an expanded version that can support more than 4,600 GPUs within a single unit, with a service launch planned within this year,” said <a href="https://www.linkedin.com/in/heonhyeock-cho-29427b147/?originalSubdomain=kr" rel="noopener noreferrer" target="_blank">Heon Hyeock Cho</a>, vice president and head of the datacenter business unit at LG CNS. LG’s first Modular Data Center will roll out in the South Korean port city of Busan, where it could deploy up to 50 units.</p><p>LG and Duos are not alone. <a href="https://www.hpe.com/us/en/services/ai-mod-pod.html" rel="noopener noreferrer" target="_blank">Hewlett Packard Enterprise,</a> <a href="https://www.vertiv.com/en-emea/solutions/vertiv-modular-solutions/?utm_source=press-release&utm_medium=public-relations&utm_campaign=hpc-ai&utm_content=en-coolchip" rel="noopener noreferrer" target="_blank">Vertiv</a>, and <a href="https://www.se.com/ww/en/work/solutions/data-centers-and-networks/modular-data-center/" rel="noopener noreferrer" target="_blank">Schneider Electric</a> now have modular data centers available or in development. A <a href="https://www.grandviewresearch.com/industry-analysis/modular-data-center-market-report" target="_blank">report</a> from market research firm <a href="https://www.grandviewresearch.com/" target="_blank">Grand View Research</a> estimates that the market for modular data centers could more than double by 2030.</p><h2>On the grid, but under the radar</h2><p>A modular data center site is quite different from traditional data center because there’s no need to construct a large steel-and-concrete shell. Instead, the site can be made ready by pouring a concrete pad. The pre-fabricated modules are delivered by truck, placed on the pad where desired, and then networked on-site.<br/><br/>Duos’ deployments, for instance, include power modules placed alongside the compute pods, and the pods are networked together with redundant fiber connections that allow the pods to operate in unison. Recker compared it to lining up school buses in a parking lot. “Everything is built off-site at a factory, and we can put it together like a jigsaw puzzle,” he said.</p><p>That simplicity is the point. Both Duos and LG CNS expect a modular data center can be deployed in about six months, compared to the roughly two or three years a conventional data center requires. Recker said that, for Duos, the turnaround is so quick that building the pre-fabricated unit isn’t always the constraint. While it’s possible to construct a pre-fabricated unit in 60 or 90 days, site preparation extends the timeline “because you can’t get the permits that fast.”</p><p>Modular data centers may also provide good value. Recker said a five-megawatt modular deployment can be built for about $25 million, and that Duos’ cost per megawatt is roughly half what larger facilities charge. For Duos, savings are possible in part because its modular data centers can target smaller deployments where the permitting is less complex. Smaller, modular deployments also meet less resistance from local governments, which are increasingly skeptical about data center construction. </p><p>While Duos targets smaller deployments, LG hopes to go big. Its planned Busan campus of 50 AI Modular Data Centers suggests an ambition to achieve deployments that rival the capacity of conventional facilities. A site with 50 units would bring the total number of GPUs to over 28,000. Here, the benefits of a modular approach could stem mostly from scalability, as a modular data center could start small and grow as required.</p><p>“By adopting a modular approach, the AI Modular Data Center can be incrementally expanded through the combination of dozens of AI Boxes,” Cho said. “It’s enabling the construction of even hyperscale-level AI data centers.”</p>]]></description><pubDate>Mon, 30 Mar 2026 14:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/modular-data-center</guid><category>Data-center</category><category>Networking</category><category>Liquid-cooling</category><category>Ai</category><dc:creator>Matthew S. Smith</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/overhead-view-of-two-data-center-pods-each-measuring-55-feet-long-by-12-5-feet-wide.jpg?id=65417343&amp;width=980"></media:content></item><item><title>Facial Recognition Is Spreading Everywhere</title><link>https://spectrum.ieee.org/facial-recognition-gone-wrong</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/illustration-34-orange-women-icons-1-blue-man-icon-labels-for-skin-tone-and-gender-comparisons.jpg?id=65407585&width=1200&height=800&coordinates=0%2C12%2C0%2C13"/><br/><br/><p>Facial recognition technology (FRT) dates back 60 years. Just over a decade ago, deep-learning methods tipped the technology into more useful—<a href="https://spectrum.ieee.org/china-facial-recognition" target="_blank">and menacing</a>—territory. Now, retailers, your neighbors, and law enforcement are all storing your face and building up a fragmentary photo album of your life.</p><p>Yet the story those photos can tell inevitably has errors. FRT makers, like those of any diagnostic technology, must balance two types of errors: false positives and false negatives. There are three possible outcomes.</p><div class="ieee-sidebar-medium"><h3>Three Possible Outcomes</h3><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="White figures and an orange hooded figure, focusing on the hooded figure in a split design." class="rm-shortcode" data-rm-shortcode-id="8a762ebf2761a791f12500ed10596cc3" data-rm-shortcode-name="rebelmouse-image" id="f4d64" loading="lazy" src="https://spectrum.ieee.org/media-library/white-figures-and-an-orange-hooded-figure-focusing-on-the-hooded-figure-in-a-split-design.png?id=65407894&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">a) identifies the suspect, since the two images are of the same person, according to the software. Success!</small></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Abstract figures: orange hoodie enlarged, white, yellow, and orange on left, black background." class="rm-shortcode" data-rm-shortcode-id="3d130b8e4c73ee49898645524cecd1f6" data-rm-shortcode-name="rebelmouse-image" id="30881" loading="lazy" src="https://spectrum.ieee.org/media-library/abstract-figures-orange-hoodie-enlarged-white-yellow-and-orange-on-left-black-background.png?id=65407867&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">b) matches another person in the footage with the suspect’s probe image. A false positive, coupled with sloppy verification, could put the wrong person behind bars and lets the real criminal escape justice.</small></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Three white icons and one orange hoodie icon on left, large orange hoodie icon on right." class="rm-shortcode" data-rm-shortcode-id="4cdaa23680c5144a5c284fcd8cb6f3df" data-rm-shortcode-name="rebelmouse-image" id="fbc8f" loading="lazy" src="https://spectrum.ieee.org/media-library/three-white-icons-and-one-orange-hoodie-icon-on-left-large-orange-hoodie-icon-on-right.png?id=65407858&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">c) fails to find a match at all. The suspect may be evading cameras, but if cameras just have low-light or bad-angle images, this creates a false negative. This type of error might let a suspect off and raise the cost of the manhunt.</small></p></div><p>In best-case scenarios—such as comparing someone’s passport photo to a photo taken by a border agent—false-negative rates are <a href="https://face.nist.gov/frte/reportcards/11/clearviewai_003.html" target="_blank">around two in 1,000 and false positives are less than one in 1 million</a>.</p><p>In the rare event you’re one of those false negatives, a border agent might ask you to show your passport and take a second look at your face. But as people ask more of the technology, more ambitious applications could lead to more catastrophic errors. Let’s say that police are searching for a suspect, and they’re comparing an image taken with a security camera with a previous “mug shot” of the suspect.</p><p>Training-data composition, differences in how sensors detect faces, and intrinsic differences between groups, such as age, all affect an algorithm’s performance. The <a href="https://assets.publishing.service.gov.uk/media/693002a4cdec734f4dff4149/1a_Cognitec_NPL_Equitability_Report_October_25.pdf&sa=D&source=docs&ust=1774557264829489&usg=AOvVaw13R0ue8NITZ-0tPVLcJ8S-" target="_blank">United Kingdom estimated</a> that its FRT exposed some groups, such as women and darker-skinned people, to risks of misidentification as high as two orders of magnitude greater than it did to others.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Five faces arranged left to right, from easy to hard to recognize." class="rm-shortcode" data-rm-shortcode-id="ce19d3eb3745de15489274ebe5083f06" data-rm-shortcode-name="rebelmouse-image" id="3ab1e" loading="lazy" src="https://spectrum.ieee.org/media-library/five-faces-arranged-left-to-right-from-easy-to-hard-to-recognize.png?id=65407777&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">Less clear photographs are harder for FRT to process.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">iStock</small></p><p>What happens with photos of people who aren’t cooperating, or vendors that train algorithms on biased datasets, or field agents who demand a swift match from a huge dataset? Here, things get murky.</p><div class="ieee-sidebar-medium"><h3>Facial Recognition Gone Wrong</h3><p><strong>THE NEGATIVES OF FALSE POSITIVES</strong></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Detroit Police SUV with American flag decal on side under bright sunlight." class="rm-shortcode" data-rm-shortcode-id="1a424f342f44dff48e8b6b05c79f5032" data-rm-shortcode-name="rebelmouse-image" id="c102c" loading="lazy" src="https://spectrum.ieee.org/media-library/detroit-police-suv-with-american-flag-decal-on-side-under-bright-sunlight.png?id=65407650&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">2020: <a href="https://quadrangle.michigan.law.umich.edu/issues/winter-2024-2025/flawed-facial-recognition-technology-leads-wrongful-arrest-and-historic&sa=D&source=docs&ust=1774557264902408&usg=AOvVaw3xUv5_o_zg1Fh0EScZ9lTW" target="_blank">Robert Williams’s wrongful arrest</a> cost him detention. The ensuing settlement requires Detroit police to enact policies that recognize FRT’s limits. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">iStock</small></p><p><strong>ALGORITHMIC BIAS</strong></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt='Red sign reads "Security cameras in use" with camera graphic.' class="rm-shortcode" data-rm-shortcode-id="014ac05f2fe587ca01643c64c750e331" data-rm-shortcode-name="rebelmouse-image" id="f4f1f" loading="lazy" src="https://spectrum.ieee.org/media-library/red-sign-reads-security-cameras-in-use-with-camera-graphic.png?id=65407620&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">2023: <a href="https://incidentdatabase.ai/cite/619/&sa=D&source=docs&ust=1774557264903427&usg=AOvVaw3fBw_78OyUB3Sa_cPpxmCi" target="_blank">Court bans Rite Aid from using facial recognition for five years</a> over its use of a racially biased algorithm. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">iStock</small></p><p><strong>TOO FAST, TOO FURIOUS?</strong></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Back of ICE officer in tactical gear facing a house." class="rm-shortcode" data-rm-shortcode-id="0004b023a075c21698cdf88cfd0b4106" data-rm-shortcode-name="rebelmouse-image" id="889f9" loading="lazy" src="https://spectrum.ieee.org/media-library/back-of-ice-officer-in-tactical-gear-facing-a-house.png?id=65407619&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">2026: U.S. immigration agents <a href="https://www.404media.co/ices-facial-recognition-app-misidentified-a-woman-twice/&sa=D&source=docs&ust=1774557264904407&usg=AOvVaw03DUrBl3YxN6c3uhHa611f" target="_blank">misidentify a woman they’d detained as two different women</a>. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">VICTOR J. BLUE/BLOOMBERG/GETTY IMAGES </small></p></div><p><span>Consider a busy trade fair using FRT to check attendees against a database, or gallery, of images of the 10,000 registrants, for example. Even at 99.9 percent accuracy you’ll get about a dozen false positives or negatives, which may be worth the trade-off to the fair organizers. But if police start using something like that across a city of 1 million people, the number of potential victims of mistaken identity rises, as do the stakes.</span></p><p><span>What if we ask FRT to tell us if the government has ever recorded and stored an image of a given person? That’s what U.S. Immigration and Customs Enforcement <a href="https://illinoisattorneygeneral.gov/News-Room/Current-News/001%20-%20Complaint%201.12.26.pdf?language_id=1" target="_blank">agents have done since June 2025</a>, using the Mobile Fortify app. The agency conducted more than 100,000 FRT searches in the first six months. The size of the potential gallery is at least <a href="https://sam.gov/opp/b016354c5bd045fa92e4886878747dc8/view" target="_blank">1.2 billion images</a>.</span></p><p><span>At that size, assuming even best-case images, the system is likely to return around 1 million false matches, but at a rate at least 10 times as high for darker-skinned people, depending on the subgroup.</span></p><p>Responsible use of this powerful technology would involve independent identity checks, multiple sources of data, and a clear understanding of the error thresholds, says computer scientist <a href="https://www.cics.umass.edu/about/directory/erik-learned-miller" target="_blank">Erik Learned-Miller</a> of the University of Massachusetts Amherst: “<a href="https://spectrum.ieee.org/joy-buolamwini" target="_blank">The care we take</a> in deploying such systems should be proportional to the stakes.”</p>]]></description><pubDate>Mon, 30 Mar 2026 13:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/facial-recognition-gone-wrong</guid><category>Facial-recognition</category><category>Privacy</category><category>Surveillance</category><category>Machine-vision</category><category>Computer-vision</category><dc:creator>Lucas Laursen</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/illustration-34-orange-women-icons-1-blue-man-icon-labels-for-skin-tone-and-gender-comparisons.jpg?id=65407585&amp;width=980"></media:content></item><item><title>How NYU’s Quantum Institute Bridges Science and Application</title><link>https://spectrum.ieee.org/nyu-quantum-institute</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/person-in-white-suit-working-with-semiconductor-equipment-in-a-lab.jpg?id=65322091&width=1200&height=800&coordinates=350%2C0%2C350%2C0"/><br/><br/><p><em>This sponsored article is brought to you by <a href="https://engineering.nyu.edu/" rel="noopener noreferrer" target="_blank">NYU Tandon School of Engineering</a>.</em></p><p>Within a 6 mile radius of New York University’s (NYU) campus, there are more than 500 tech industry giants, banks, and hospitals. This isn’t just a fact about real estate, it’s the foundation for advancing quantum discovery and application.</p><p>While the world races to harness quantum technology, NYU is betting that the ultimate advantage lies not solely in a lab, but in the dense, demanding, and hyper-connected urban ecosystem that surrounds it. With the launch of its <a href="https://www.nyu.edu/about/news-publications/news/2025/october/nyu-launches-quantum-institute-.html" rel="noopener noreferrer" target="_blank"><span>NYU Quantum Institute</span></a> (NYUQI), NYU is positioning itself as <a href="https://www.nyu.edu/about/news-publications/news/2025/october/top-quantum-scientists-convene-at-nyu.html" target="_blank">the central node</a> in this network; a “full stack” powerhouse built on the conviction that it has found the right place, and the right time, to turn quantum science into tangible reality.</p><p>Proximity advantage is essential because quantum science demands it. Globally, the quest for practical quantum solutions — whether for computing, sensing, or secure communications — has been stalled, in part, by fragmentation. Physicists and chemical engineers invent new materials, computer scientists develop new algorithms, and electrical engineers build new devices, but all three often work in isolated academic silos.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Three men pose at the 4th Annual NYC Quantum Summit 2025; attendees converse in the background." class="rm-shortcode" data-rm-shortcode-id="1dd6dfe45b73630bb9040545fcdfae7d" data-rm-shortcode-name="rebelmouse-image" id="33e2d" loading="lazy" src="https://spectrum.ieee.org/media-library/three-men-pose-at-the-4th-annual-nyc-quantum-summit-2025-attendees-converse-in-the-background.jpg?id=65322345&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Gregory Gabadadze, NYU’s dean for science, NYU physicist and Quantum Institute Director Javad Shabani, and Juan de Pablo, Anne and Joel Ehrenkranz Executive Vice President for Global Science and Technology and executive dean of the Tandon School of Engineering.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Veselin Cuparić/NYU</small></p><p><span>NYUQI’s premise is that breakthroughs happen “at the interfaces between different domains,” according to </span><a href="https://engineering.nyu.edu/faculty/juan-de-pablo" target="_blank"><span>Juan de Pablo</span></a><span>, Executive Vice President for Global Science and Technology at NYU and Executive Dean of the NYU Tandon School of Engineering. The Institute is built to actively force those necessary collisions — to integrate the physicists, engineers, materials scientists, computer scientists, biologists, and chemists vital to quantum research into one holistic operation. This institutional design ensures that the hardware built by one team can be immediately tested by software developed by another, accelerating progress in a way that isolated departments never could.</span></p><p class="pull-quote"><span>NYUQI’s premise is that breakthroughs happen at the interfaces between different domains. <strong>—Juan de Pablo, NYU Tandon School of Engineering</strong></span></p><p>NYUQI’s integrated vision is backed by a massive physical commitment to the city. The NYUQI is not just a theoretical concept; its collaborators will be housed in a renovated, <a href="https://www.nyu.edu/about/news-publications/news/2025/may/nyu-entering-long-term-lease-at-770-broadway.html" target="_blank"><span>million-square-foot facility</span></a> in the heart of Manhattan’s West Village, backed by a state-of-the-art <a href="https://engineering.nyu.edu/research/nanofab" target="_blank">Nanofabrication Cleanroom</a> in Brooklyn serving as a high-tech foundry. This is where the theoretical meets physical devices, allowing the Institute to test and refine the process from materials science to deployment.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt='NYU building exterior with "Science + Tech" signage, flags, and a passing yellow taxi.' class="rm-shortcode" data-rm-shortcode-id="605cc71d844927d3fb0a05fb086fedcf" data-rm-shortcode-name="rebelmouse-image" id="bceaa" loading="lazy" src="https://spectrum.ieee.org/media-library/nyu-building-exterior-with-science-tech-signage-flags-and-a-passing-yellow-taxi.jpg?id=65322352&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">NYUQI will be housed in a renovated, million-square-foot facility in the heart of Manhattan’s West Village.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Tracey Friedman/NYU</small></p><p><span>Leading this effort is NYUQI Director </span><a href="https://as.nyu.edu/faculty/javad-shabani.html" target="_blank"><span>Javad Shabani</span></a><span>, who, along with the other members, is turning the Institute into a hub for collaboration with private and public sector partners with quantum challenges that need solving. As de Pablo explains, “Anybody who wants to work on quantum with NYU, you come in through that door, and we’ll send you to the right place.” For New York’s vast ecosystem of tech giants and financial institutions, the NYUQI offers a resource they can’t build on their own: a cohesive team of experts in quantum phenomena, quantum information theory, communication, computing, materials, and optics, and a structured path to applying theoretical discoveries to advanced quantum technologies.</span></p><h2>Solving the Challenge of Quantum Research</h2><p><span>The NYUQI’s integrated structure is less about organizational management, and more about scientific requirement. </span><span>The challenge of quantum is that the hardware, the software, and the programming are inherently interconnected — each must be designed to work with the other. To solve this, the Institute focuses on three applications of quantum science: Quantum Computing, Quantum Sensing, and Quantum Communications.</span></p><p>For Shabani, this means creating an integrated environment that bridges discovery with experimentation, starting with the physical components all the way to quantum algorithm centers. That will include a fabrication facility in the new building in Manhattan, as well as the <a href="https://engineering.nyu.edu/news/chips-and-science-act-spurs-nanofab-cleanroom-ribbon-cutting-nyu-tandon-school-engineering" target="_blank"><span>NYU Nanofab</span></a> in Brooklyn directed by Davood Shahjerdi. New York Senators Charles Schumer and Kirsten Gillibrand recently secured <a href="https://www.nyu.edu/about/news-publications/news/2026/february/nyu-receives--1-million-in-funding-from-senators-schumer-and-gil.html" target="_blank">$1 million in congressionally-directed spending</a> to bring Thermal Laser Epitaxy (TLE) technology — which allows for atomic-level purity, minimal defects, and streamlined application of a diverse range of quantum materials — to NYU, marking the first time the equipment will be used in the U.S.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Two people hold semiconductor wafers during a presentation with audience taking photos." class="rm-shortcode" data-rm-shortcode-id="1a0dbca6c6bb8fb7dbf4d399689b2922" data-rm-shortcode-name="rebelmouse-image" id="d434c" loading="lazy" src="https://spectrum.ieee.org/media-library/two-people-hold-semiconductor-wafers-during-a-presentation-with-audience-taking-photos.jpg?id=65322354&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">NYU Nanofab manager Smiti Bhattacharya and Nanofab Director Davood Shahjerdi at the nanofab ribbon-cutting in 2023. The nanofab is the first academic cleanroom in Brooklyn, and serves as a prototyping facility for the NORDTECH Microelectronics Commons consortium.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">NYU WIRELESS</small></p><p>Tight control over fabrication, and can allow researchers to pivot quickly when a breakthrough in one area — say, finding a cheaper, more reliable material like silicon carbide — can be explored for use across all three applications, and offers unique access to academics and the private sector alike to sophisticated pieces of specialty equipment whose maintenance knowledge and costs make them all-but-impossible to maintain outside of the right staffing and environment.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="3D model of a laboratory layout, highlighting the Yellow Room in bright yellow." class="rm-shortcode" data-rm-shortcode-id="e7c1128703d96de919ed2ce440a97416" data-rm-shortcode-name="rebelmouse-image" id="62d58" loading="lazy" src="https://spectrum.ieee.org/media-library/3d-model-of-a-laboratory-layout-highlighting-the-yellow-room-in-bright-yellow.png?id=65322596&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">The NYU Nanofab is Brooklyn’s first academic cleanroom, with a strategic focus on superconducting quantum technologies, advanced semiconductor electronics, and devices built from quantum heterostructures and other next-generation materials.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">NYU Nanofab</small></p><p><span>That speed and adaptability is the NYUQI’s competitive edge. It turns fragmented challenges into holistic solutions, positioning the Institute to solve real-world problems for its New York neighbors—from highly secure data transmission to next-generation drug discovery.</span></p><h2>Testing Quantum Communication in NYC</h2><p>The integrated approach also makes the NYUQI a testbed for the most critical near-term applications. Take Quantum Communications, which is essential for creating an “unhackable” quantum internet. In an industry first, NYU worked with the quantum start-up Qunnect to <a href="https://www.nyu.edu/about/news-publications/news/2023/september/nyu-takes-quantum-step-in-establishing-cutting-edge-tech-hub-in-.html" target="_blank"><span>send quantum information through standard telecom fiber</span></a> in New York City between Manhattan and Brooklyn through a 10-mile quantum networking link. Instead of simulating communication challenges in a lab, the NYUQI team is already leveraging NYU’s city-wide campus by utilizing existing infrastructure to test secure quantum transmission between Manhattan and Brooklyn. </p><p class="pull-quote">The NYUQI team is already leveraging NYU’s city-wide campus by utilizing existing infrastructure to test secure quantum transmission between Manhattan and Brooklyn.</p><p>This isn’t just theory; it is building a functioning prototype in the most demanding, dense urban environment  in the world. Real-time, real-world deployment is a critical component missing in other isolated institutions. When the NYUQI achieves results, the technology will be that much more readily available to the massive financial, tech, and communications organizations operating right outside their door.</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="Scientist in protective gear working in a laboratory with samples." class="rm-shortcode" data-rm-shortcode-id="d644b791788af64769a853d0516834e6" data-rm-shortcode-name="rebelmouse-image" id="dc2fb" loading="lazy" src="https://spectrum.ieee.org/media-library/scientist-in-protective-gear-working-in-a-laboratory-with-samples.jpg?id=65322378&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">NYUQI includes a state-of-the-art Nanofabrication Cleanroom in Brooklyn serving as a high-tech foundry.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">NYU Tandon</small></p><p><span>While the Institute has built the physical infrastructure and designed the necessary scientific architecture, its enduring contribution will be the specialized workforce it creates for the new quantum economy. This addresses the market’s greatest deficit: a lack of individuals trained not just in physics, but in the integrated, full-stack approach that quantum demands.</span></p><p>By creating a pipeline of 100 to 200 graduate and doctoral students who are encouraged to collaborate across Computing, Sensing, and Communications, the NYUQI is narrowing the skills gap. These will be future leaders who can speak the language of the physicist, the materials scientist, and the engineer simultaneously. This commitment to interdisciplinary talent is also fueled by the launch of the new Master of Science in Quantum Science & Technology program at NYU Tandon, positioning the university among a select group worldwide offering such a specialized degree.</p><p>Interdisciplinary education creates the shared language and understanding poised to make graduates coming from collaborations in the NYUQI extremely valuable in the current landscape. Quantum challenges are not just technical; they are managerial and philosophical as well. An engineer working with the NYUQI will understand the requirements of the nanofabrication cleanroom and the foundations of superconducting qubits for quantum computing, just as a physicist will understand the application needs of an industry partner like a large financial institution. In a field where the entire team must be able to communicate seamlessly, these are professionals truly equipped to rapidly translate discovery into deployable technology. Creating a talent pipeline at scale will provide a missing link that converts New York’s vast commercial energy into genuine quantum advantage.</p><h2>NYUQI: Building Talent, Technology, and Structure</h2><p><span>The vision for the NYUQI </span><span>is an act of strategic geography that plays directly into the sheer volume of opportunity and demand right outside their new facility. </span><span>By building the talent, the technology, and the structure necessary to capitalize on this dense environment, NYU is not just participating in the quantum race, it is actively steering it.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Conference room with attendees seated at round tables, facing a presenter on stage." class="rm-shortcode" data-rm-shortcode-id="f5e2ae16e0c5ebc4f0828d52ed639115" data-rm-shortcode-name="rebelmouse-image" id="02b7e" loading="lazy" src="https://spectrum.ieee.org/media-library/conference-room-with-attendees-seated-at-round-tables-facing-a-presenter-on-stage.jpg?id=65322370&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Attendees of NYU’s 2025 Quantum Summit.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Tracey Friedman/NYU</small></p><p>The initial hypothesis for the NYUQI was simple: the ultimate advantage lies in pursuing the science in the right place at the right time. Now, the institute will ensure that the next wave of scientific discovery, capable of solving previously intractable problems in finance, medicine, and security, will be conceived, built, and tested in the heart of New York City.</p>]]></description><pubDate>Fri, 27 Mar 2026 10:02:05 +0000</pubDate><guid>https://spectrum.ieee.org/nyu-quantum-institute</guid><category>Nyu-tandon</category><category>Quantum-computing</category><category>Quantum-internet</category><category>Semiconductors</category><category>Quantum-communications</category><dc:creator>Wiley</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/person-in-white-suit-working-with-semiconductor-equipment-in-a-lab.jpg?id=65322091&amp;width=980"></media:content></item><item><title>How IEEE 802.11bn Delivers Ultra-High Reliability for Wi-Fi 8</title><link>https://content.knowledgehub.wiley.com/setting-new-performance-standards-with-ieee-802-11bn-an-in-depth-overview-of-wi-fi-8/</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/logo-of-rohde-schwarz-with-slogan-make-ideas-real-and-stylized-rs-in-a-diamond-shape.png?id=65355284&width=980"/><br/><br/><p><span>A technical exploration of IEEE 802.11bn’s physical and MAC layer enhancements — including distributed resource units, enhanced long range, multi-AP coordination, and seamless roaming — that define Wi-Fi 8.</span></p><p><strong><span>What Attendees will Learn</span></strong></p><ol><li><span>Why Wi-Fi 8 prioritizes reliability over raw throughput — Understand how IEEE 802.11bn shifts the design philosophy from peak data-rate gains to ultra-high reliability.</span></li><li>How new physical layer features overcome uplink power limitations — Learn how distributed resource units spread tones across wider distribution bandwidths to boost per-tone transmit power, and how enhanced long range protocol data units use power-boosted preamble fields and frequency-domain duplication to extend uplink coverage.</li><li>How advanced MAC coordination reduces interference and latency — Examine multi-access point coordination schemes — coordinated beamforming, spatial reuse, time division multiple access, and restricted target wake time — alongside non-primary channel access and priority enhanced distributed channel access.</li><li>What seamless roaming and power management mean for next-generation deployments — Discover how seamless mobility domains eliminate reassociation delays during access point transitions, and how dynamic power save and multi-link power management let devices trade capability for battery life without sacrificing connectivity.</li></ol><p><a href="https://content.knowledgehub.wiley.com/setting-new-performance-standards-with-ieee-802-11bn-an-in-depth-overview-of-wi-fi-8/" target="_blank">Download this free whitepaper now!</a></p>]]></description><pubDate>Wed, 25 Mar 2026 14:22:07 +0000</pubDate><guid>https://content.knowledgehub.wiley.com/setting-new-performance-standards-with-ieee-802-11bn-an-in-depth-overview-of-wi-fi-8/</guid><category>Wifi</category><category>Internet</category><category>Standards</category><category>Transmission</category><category>Type-whitepaper</category><dc:creator>Rohde &amp; Schwarz</dc:creator><media:content medium="image" type="image/png" url="https://assets.rbl.ms/65355284/origin.png"></media:content></item><item><title>Data Centers Are Transitioning From AC to DC</title><link>https://spectrum.ieee.org/data-center-dc</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/nvidia-s-high-compute-density-racks.jpg?id=65397940&width=1200&height=800&coordinates=0%2C208%2C0%2C209"/><br/><br/><p>Last week’s <a href="https://www.nvidia.com/gtc/" target="_blank">Nvidia GTC</a> conference highlighted new <a href="https://spectrum.ieee.org/nvidia-groq-3" target="_blank">chip</a> architectures to power AI. But as the chips become faster and more powerful, the remainder of data center <a data-linked-post="2674166715" href="https://spectrum.ieee.org/data-center-liquid-cooling" target="_blank">infrastructure</a> is playing catch-up. The power-delivery community  is responding: Announcements from <a href="https://www.prnewswire.com/news-releases/delta-exhibits-energy-saving-solutions-for-800-vdc-in-next-gen-ai-factories-and-digital-twin-applications-built-on-omniverse-at-nvidia-gtc-2026-302715850.html" rel="noopener noreferrer" target="_blank">Delta</a>,  <a href="https://www.eaton.com/us/en-us/company/news-insights/news-releases/2026/eaton-collaborates-with-nvidia-to-unveil-its-beam-rubin-dsx-platform.html" rel="noopener noreferrer" target="_blank">Eaton</a>, <a href="https://www.se.com/us/en/about-us/newsroom/news/press-releases/Schneider-Electric-teams-with-NVIDIA-to-develop-validated-blueprints-to-design-simulate-build-operate-and-maintain-gigawattscale-AI-Factories-69b82f61aa1027e04205d273/" target="_blank">Schneider Electric</a>, and <a href="https://www.vertiv.com/en-us/about/news-and-insights/corporate-news/2026/vertiv-brings-converged-physical-infrastructure-to-nvidia-vera-rubin-dsx-ai-factories/" rel="noopener noreferrer" target="_blank">Vertiv</a> showcased new designs for the AI era. Complex and inefficient AC-to-DC power conversions are gradually being replaced by DC configurations, at least in hyperscale data centers.</p><p>“While AC distribution remains deeply entrenched, advances in power electronics and the rising demands of AI infrastructure are accelerating interest in DC architectures,” says <a href="https://www.linkedin.com/in/solarchris/" target="_blank">Chris Thompson</a>, vice president of advanced technology and global microgrids at Vertiv.</p><h2>AC-to-DC Conversion Challenges</h2><p>Today, nearly all data centers are designed around AC utility power. The electrical path includes multiple conversions before power reaches the compute load. Power typically enters the data center as medium-voltage AC (1 to 35 kilovolts), is stepped down to low-voltage AC (480 or 415 volts) using a transformer, converted to DC inside an uninterruptible power supply (UPS) for battery storage, converted back to AC, and converted again to low-voltage DC (typically 54 V DC) at the server, supplying the DC power computing chips actually require.</p><p>“The double conversion process ensures the output AC is clean, stable, and suitable for data center servers,” says <a href="https://www.linkedin.com/in/luiz-fernando-huet-de-bacellar-b2112117/" target="_blank">Luiz Fernando Huet de Bacellar,</a> vice president of engineering and technology at Eaton.</p><p>That setup worked well enough for the amounts of power required by traditional data centers. Traditional data center computational racks draw on the order of 10 kW each. For AI, that is starting to approach 1 megawatt.  At that scale, the energy losses, current levels, and copper requirements of AC-to-DC conversions become increasingly difficult to justify. Every conversion incurs some power loss. On top of that, as the amount of power that needs to be delivered grows, the sheer size of the convertors, as well as the connector requirements of copper busbars, becomes untenable.<span> According to an Nvidia <a href="https://developer.nvidia.com/blog/nvidia-800-v-hvdc-architecture-will-power-the-next-generation-of-ai-factories/" target="_blank">blog</a>, a 1-MW rack</span><span> could require as much as 200 kilograms of copper busbar. For a 1-gigawatt data center, it could amount to 200,000 kg of copper. </span></p><h2>Benefits of High-Voltage DC Power</h2><p>By converting 13.8-kV AC grid power directly to 800 V DC at the data center perimeter, most intermediate conversion steps are eliminated. This reduces the number of fans and power-supply units, and leads to higher system reliability, lower heat dissipation, improved energy efficiency, and a smaller equipment footprint.</p><p>“Each power conversion between the electric grid or power source and the silicon chips inside the servers causes some energy loss,” says Bacellar.</p><p>Switching from 415-V AC to 800-V DC in electrical distribution enables 85 percent more power to be transmitted through the same conductor size. This happens because higher voltage reduces current demand, lowering resistive losses and making power transfer more efficient. Thinner conductors can handle the same load, reducing copper requirements by 45 percent, a 5 percent improvement in efficiency, and 30 percent lower total cost of ownership for gigawatt-scale facilities.</p><p>“In a high-voltage DC architecture, power from the grid is converted from medium-voltage AC to roughly 800-V DC and then distributed throughout the facility on a DC bus,” said Vertiv’s Thompson. “At the rack, compact DC-to-DC converters step that voltage down for GPUs and CPUs.”</p><p>A <a href="https://www.datacenter-asia.com/wp-content/uploads/2025/08/Omdia-Analysts-Summit-Omdia%E5%88%86%E6%9E%90%E5%B8%88%E5%B3%B0%E4%BC%9A.pdf" target="_blank">report</a> from technology advisory group <a href="https://omdia.tech.informa.com/" target="_blank">Omdia</a> claims that higher voltage DC data centers have already appeared in China. In the Americas, the <a href="https://www.linkedin.com/posts/sharada-yeluri_microsoft-meta-google-activity-7367974455052017666-nXV5/" target="_blank">Mt. Diablo Initiative</a> (a collaboration among <a href="https://www.meta.com/about/?srsltid=AfmBOoq7uBjCU2oG3oI6Ti8VQaMdaxhAcxXmXD-twy9OTi0cbmTqGKVQ" target="_blank">Meta</a>, <a href="https://www.microsoft.com/en-us" target="_blank">Microsoft</a>, and the <a href="https://www.opencompute.org/" target="_blank">Open Compute Project</a>) is a 400-V DC rack power distribution experiment.</p><h2>Innovations in DC Power Systems</h2><p>A handful of vendors are trying to get ahead of the game. Vertiv’s 800-V DC ecosystem that integrates with <a href="https://www.vertiv.com/en-us/about/news-and-insights/corporate-news/vertiv-develops-energy-efficient-cooling-and-power-reference-architecture-for-the-nvidia-gb300-nvl72/" target="_blank">Nvidia Vera Rubin Ultra Kyber platforms</a> will be commercially available in the second half of 2026. Eaton, too, is well advanced in its 800-V DC systems innovation courtesy of a medium-voltage solid-state transformer (SST) that will sit at the heart of DC power distribution system. Meanwhile Delta, has released 800-V DC in-row 660-kW power racks with a total of 480 kW of embedded battery backup units. And, <a href="https://www.solaredge.com/us/" target="_blank">SolarEdge</a> is hard at work on a 99%-efficient SST that will be paired with a native DC UPS and a DC power distribution layer.</p><p>But much of the industry is far behind. <a href="https://www.linkedin.com/in/pehughes/" target="_blank">Patrick Hughes</a>, senior vice president of strategy, technical, and industry affairs for the <a href="https://www.makeitelectric.org/" target="_blank">National Electrical Manufacturers Association</a>, says most innovation is happening at the 400-V DC level, though some are preparing 800-V DC. He believes the industry needs a complete, coordinated ecosystem, including power electronics, protection, connectors, sensing, and service‑safe components that scale together rather than in isolation. That, in turn, requires retooling manufacturing capacity for DC‑specific equipment, expanding semiconductor and materials supply, and clear, long‑term demand commitments that justify major capital investment across the value chain.</p><p>“Many are taking a cautious approach, offering limited or adapted solutions while waiting for clearer standards, safety frameworks, and customer commitments,” said Hughes. “Building the supply chain will hinge on stabilizing standards and safety frameworks so suppliers can design, certify, manufacture, and install equipment with confidence.”</p>]]></description><pubDate>Tue, 24 Mar 2026 16:00:05 +0000</pubDate><guid>https://spectrum.ieee.org/data-center-dc</guid><category>Data-centers</category><category>Power-electronics</category><category>Ai</category><dc:creator>Drew Robb</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/nvidia-s-high-compute-density-racks.jpg?id=65397940&amp;width=980"></media:content></item><item><title>What Will It Take to Build the World’s Largest Data Center?</title><link>https://spectrum.ieee.org/5gw-data-center</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/construction-symbols-on-yellow-background.png?id=65356154&width=1200&height=800&coordinates=0%2C869%2C0%2C870"/><br/><br/><p><strong>The undying thirst for </strong>smarter (historically, that means larger) AI models and greater adoption of the ones we already have has led to an explosion in <a href="https://epoch.ai/data/data-centers#data-insights" rel="noopener noreferrer" target="_blank">data-center construction projects</a>, unparalleled both in number and scale. Chief among them is Meta’s planned 5-gigawatt data center in Louisiana, called Hyperion, announced in June of 2025. Meta CEO Mark Zuckerberg said Hyperion will “cover a significant part of the footprint of Manhattan,” and the first phase—a 2-GW version—will be completed by 2030.</p><p>Though the project’s stated 5-GW scale is the largest among its peers, it’s just one of several dozen similar projects now underway. According to Michael Guckes, chief economist at construction-software company <a href="https://www.constructconnect.com/preconstruction-software?campaign=21011210878&group=161161401080&target=kwd-337013613104&matchtype=e&creative=760058507701&device=c&se_kw=constructconnect&utm_medium=ppc&utm_campaign=CC+Brand+2&utm_term=constructconnect&utm_source=adwords&hsa_ad=760058507701&hsa_kw=constructconnect&hsa_net=adwords&hsa_tgt=kwd-337013613104&hsa_grp=161161401080&hsa_src=g&hsa_ver=3&hsa_cam=21011210878&hsa_mt=e&hsa_acc=3324869874&gad_source=1&gad_campaignid=21011210878&gbraid=0AAAAADccs_biRlt8tR8-qu3h7Kja1Tzte&gclid=CjwKCAiA3-3KBhBiEiwA2x7FdCQc4sQOa0YZVFnCW9RF1tGkH2hDiowNrjM587XsXAv6Fb7Sdr1hgBoCNjEQAvD_BwE" rel="noopener noreferrer" target="_blank">ConstructConnect</a>, spending on data centers topped US $27 billion by July of 2025 and, once the full-year figures are tallied, will easily exceed $60 billion. Hyperion alone accounts for about a quarter of that.</p><p>For the engineers assigned to bring these projects to life, the mix of challenges involved represent a unique moment. The world’s largest tech companies are opening their wallets to pay for new innovations in compute, cooling, and <a data-linked-post="2674861846" href="https://spectrum.ieee.org/nvidia-rubin-networking" target="_blank">network</a> technology designed to operate at a scale that would’ve seemed absurd five years ago.</p><p>At the same time, the breakneck pace of building comes paired with serious problems. Modern data-center construction frequently requires an influx of temporary workers and sharply increases noise, traffic, pollution, and often local electricity prices. And the environmental toll remains a concern long after facilities are built due to the unprecedented 24/7 energy demands of AI data centers which, according to one recent study, <a href="https://www.nature.com/articles/s41893-025-01681-y" rel="noopener noreferrer" target="_blank">could emit the equivalent of tens of millions of tonnes of CO<span><sub>2</sub></span> annually</a> in the United States alone.</p><p>Regardless of these issues, large AI companies, and the engineers they hire, are going full steam ahead on giant data-center construction. So, what does it really take to build an unprecedentedly large data center?</p><h2>AI Rewrites Building Design</h2><p>The stereotypical data-center building rests on a reinforced concrete slab foundation. That’s paired with a steel skeleton and poured concrete wall panels. The finished building is called a “shell,” a term that implies the structure itself is a secondary concern. Meta has <a href="https://www.datacenterdynamics.com/en/news/meta-brings-data-centers-in-tents-to-gallatin-tennessee/" target="_blank">even used gigantic tents</a> to throw up temporary data centers.</p><p>Still, the scale of the largest AI data centers brings unique challenges. “The biggest challenge is often what’s under the surface. Unstable, corrosive, or expansive soils can lead to delays and require serious intervention,” says <a href="https://www.jacobs.com/our-people/meet-bob-haley" target="_blank">Robert Haley</a>, vice president at construction consulting firm <a href="https://www.jacobs.com/" target="_blank">Jacobs</a>.<a href="https://www.stantec.com/en/people/c/carter-amanda" target="_blank"> Amanda Carter</a>, a senior technical lead at <a href="https://www.stantec.com/en" target="_blank">Stantec</a>, said a soil’s thermal conductivity is also important, as most electrical infrastructure is placed underground. “If the soil has high thermal resistivity, it’s going to be difficult to dissipate [heat].” Engineers may take hundreds or thousands of soil samples before construction can begin.</p><h3>GPUs</h3><br/><img alt="Yellow microchip icon on a black background." class="rm-shortcode" data-rm-shortcode-id="9612db5baec52cce6fe11d703e52c7bc" data-rm-shortcode-name="rebelmouse-image" id="af54d" loading="lazy" src="https://spectrum.ieee.org/media-library/yellow-microchip-icon-on-a-black-background.png?id=65347639&width=980"/><p>Modern AI data centers often use <em><em>rack-scale</em></em> systems, such as the Nvidia GB200 NVL72, which occupy a single data-center rack. Each rack contains 72 GPUs, 36 CPUs, and up to 13.4 terabytes of GPU memory. The racks measure over 2.2 meters tall and weigh over one and a half tonnes, forcing AI data centers to use thicker concrete with more reinforcement to bear the load.</p><p>A single GB200 rack can use up to 120 kilowatts. If Hyperion meets its 5-gigawatt goals, the data-center campus could include over 41,000 rack-scale systems, for a total of more than 3 million GPUs. The final number of GPUs used by Hyperion is likely to be less than that, though only because future GPUs will be larger, more capable, and use more power.</p><h3>Money</h3><br/><img alt="Black hand and dollar symbol combined on an orange background." class="rm-shortcode" data-rm-shortcode-id="2ef34f3679a3b3135244243e46ae5630" data-rm-shortcode-name="rebelmouse-image" id="248eb" loading="lazy" src="https://spectrum.ieee.org/media-library/black-hand-and-dollar-symbol-combined-on-an-orange-background.png?id=65347751&width=980"/><p>According to ConstructConnect, spending on data centers neared US $27 billion through July of 2025 and, according to the latest data, will tally close to $60 billion through the end of the year. Meta’s Hyperion project is a big slice of the pie, at $10 billion.</p><p>Data-center spending has become an important prop for the construction industry, which is seeing reduced demand in other areas, such as residential construction and public infrastructure. ConstructConnect’s third quarter 2025 financial report stated that the quarter’s decline “would have been far more severe without an $11 billion surge in data center starts.”</p><h3></h3><br/><p>There’s apparently no shortage of eligible sites, however, as both the number of data centers under construction, and the money spent on them, has skyrocketed. The spending has allowed companies building data centers to throw out the rule book. Prior to the AI boom, most data centers relied on tried-and-true designs that prioritized inexpensive and efficient construction. Big tech’s willingness to spend has shifted the focus to speed and scale.</p><p>The loose purse strings open the door to larger and more robust prefabricated concrete wall and floor panels. <a href="https://www.linkedin.com/in/dougbevier/" target="_blank">Doug Bevier</a>, director of development at <a href="https://www.clarkpacific.com/" rel="noopener noreferrer" target="_blank">Clark Pacific</a>, says some concrete floor panels may now span up to 23 meters and need to handle floor loads up to 3,000 kilograms per square meter, <a href="https://codes.iccsafe.org/s/IBC2018/chapter-16-structural-design/IBC2018-Ch16-Sec1607.1" rel="noopener noreferrer" target="_blank">which is more than twice the load international building codes normally define for manufacturing and industry</a>. In some cases, the concrete panels must be custom-made for a project, an expensive step that the economics of pre-AI data centers rarely justified.</p><p>Simultaneously, the time scale for projects is also compressed: <a href="https://www.linkedin.com/in/jamiemcgrath365/" rel="noopener noreferrer" target="_blank">Jamie McGrath</a>, senior vice president of data-center operations at<a href="https://www.crusoe.ai/" rel="noopener noreferrer" target="_blank"> Crusoe</a>, says the company is delivering projects in “about 12 months,” compared to 30 to 36 months before. Not all projects are proceeding at that pace, but speed is universally a priority.</p><p>That makes it difficult to coordinate the labor and materials required. Meta’s Hyperion site, located in rural Richland Parish, Louisiana, is emblematic of this challenge. <a href="https://www.nola.com/news/business/meta-louisiana-ai-data-center/article_77f553ff-c272-4e6c-a775-60bbbee0b065.html" rel="noopener noreferrer" target="_blank">As reported by NOLA.com</a>, at least 5,000 temporary workers have flocked to the area, which has only about 20,000 permanent residents. These <a href="https://www.wsj.com/business/data-centers-are-a-gold-rush-for-construction-workers-6e3c5ce0?st=jr1y94" rel="noopener noreferrer" target="_blank">workers earn above-average wages</a> and bring a short-term boost for some local businesses, such as restaurants and convenience stores. However, they have also spurred complaints from residents about traffic and construction noise and pollution.</p><p>This friction with residents includes not only these obvious impacts, but <a href="https://youtu.be/DGjj7wDYaiI?si=aZocXHJe0IYUkJcl&t=175" rel="noopener noreferrer" target="_blank">also things you might not immediately suspect</a>, such as light pollution caused by around-the-clock schedules. Also significant are changes to local water tables and runoff, which can reduce water quality for neighbors who rely on well water. These issues have motivated a few U.S. cities <a href="https://www.atlantanewsfirst.com/2025/06/04/atlanta-tightens-restrictions-data-centers-bans-them-some-neighborhoods/" rel="noopener noreferrer" target="_blank">to enact data-center bans</a>.</p><h2>Data Centers Often Go BYOP (bring your own power)</h2><p>Meta’s Richland Parish site also highlights a problem that’s priority No. 1 for both AI data centers and their critics: power.</p><p>Data centers have always drawn large amounts of power, which nudged data-center construction to cluster in hubs where local utilities were responsive to their demands. Virginia’s electric utility, Dominion Energy, met demand with agreements to build new infrastructure, <a href="https://rmi.org/amazon-dominion-virginia-power-reach-breakthrough-renewable-energy-agreement/" rel="noopener noreferrer" target="_blank">often with a focus on renewable energy</a>.</p><p>The power demands of the largest AI data centers, though, have caught even the most responsive utilities off guard. A report from the Lawrence Berkeley National Laboratory, in California, estimated the entire U.S. data-center industry <a href="https://eta-publications.lbl.gov/sites/default/files/lbnl-1005775_v2.pdf" rel="noopener noreferrer" target="_blank">consumed an average load of roughly 8 GW of power in 2014</a>. Today, the largest AI data-center campuses are built to handle up to a gigawatt each, and Meta’s Hyperion is projected to require 5 GW.</p><p>“Data centers are exasperating issues for a lot of utilities,” says <a href="https://www.cleanegroup.org/staff/abbe-ramanan/" rel="noopener noreferrer" target="_blank">Abbe Ramanan</a>, project director at the Clean Energy Group, a Vermont-based nonprofit.</p><p>Ramanan explains that utilities often use “peaker plants” to cope with extra demand. They’re usually older, less efficient fossil-fuel plants which, because of their high cost to operate and carbon output, were due for retirement. But Ramanan says increased electricity demand <a href="https://www.eia.gov/todayinenergy/detail.php?id=61425" rel="noopener noreferrer" target="_blank">has kept them in service</a>.</p><p>Meta secured power for Hyperion by negotiating with Entergy, Louisiana’s electric utility, for construction of three new gas-turbine power plants. Two will be located near the Richland Parish site, while a third will be located in southeast Louisiana.</p><p>Entergy frames the new plants as a win for the state. “A core pillar of Entergy and Meta’s agreement is that Meta pays for the full cost of the utility infrastructure,” says <a href="https://www.linkedin.com/in/daniel-kline-068356ba/" rel="noopener noreferrer" target="_blank">Daniel Kline</a>, director of power-delivery planning and policy at Entergy. The utility expects that “customer bills will be lower than they otherwise would have been.” That would prove an exception, as <a href="https://www.bloomberg.com/graphics/2025-ai-data-centers-electricity-prices/?embedded-checkout=true" rel="noopener noreferrer" target="_blank">a recent report from Bloomberg found</a> electricity rates in regions with data centers are more likely to increase than in regions without.</p><h3>CO2</h3><br/><img alt="Diagram of CO2 molecule with black carbon and red oxygen atoms connected by lines." class="rm-shortcode" data-rm-shortcode-id="c9cf38ac7004d413b7fe5b8b577a3d3d" data-rm-shortcode-name="rebelmouse-image" id="3b1b0" loading="lazy" src="https://spectrum.ieee.org/media-library/diagram-of-co2-molecule-with-black-carbon-and-red-oxygen-atoms-connected-by-lines.png?id=65348689&width=980"/><p>Research <a href="https://www.nature.com/articles/s41893-025-01681-y" target="_blank">published in Nature</a> in 2025 projects that data-center emissions will range from 24 million to 44 million CO2-equivalent metric tonnes annually through 2030 in the United States alone. While some materials used in data centers, such as concrete, lead to significant emissions, the majority of these emissions will result from the high energy demands of AI servers.</p><p>Estimating the carbon emissions of Hyperion is difficult, as the project won’t be completed until 2030. Assuming that the three new natural gas plants that are planned for construction as part of the project produce emissions typical for their type, however, the plants could lead to full life-cycle emissions of between 4 million and 10 million metric tons of CO2 annually—roughly equivalent to the annual emissions of a country like <a href="https://www.worldometers.info/co2-emissions/co2-emissions-by-country/" target="_blank">Latvia</a>.</p><h3>Concrete</h3><br/><img alt="Silhouette of a cement truck on an orange background." class="rm-shortcode" data-rm-shortcode-id="060b1cd238b9de45274d6766069f3a14" data-rm-shortcode-name="rebelmouse-image" id="e6d68" loading="lazy" src="https://spectrum.ieee.org/media-library/silhouette-of-a-cement-truck-on-an-orange-background.png?id=65348696&width=980"/><p>Data centers are typically built from concrete, with steel used as a skeleton to reinforce and shape the concrete shell. While the foundation is often poured concrete, the walls and floors are most often built from prefabricated concrete panels that can span up to 23 meters. Floors use a reinforced T-shape, similar to a steel girder, measuring up to 1.2 meters across at its thickest point. The largest data centers include hundreds of these concrete panels.</p><p>The America Cement Association projects that the current surge in building<a href="https://mi.cement.org/PDF/Data_Center_Cement_Consumption.pdf" rel="noopener noreferrer" target="_blank"> will require 1 million tonnes of cement over the next three years</a>, though that’s still a tiny fraction of the overall cement industry,<a href="https://d9-wret.s3.us-west-2.amazonaws.com/assets/palladium/production/s3fs-public/media/files/mis-202507-cemen.pdf" rel="noopener noreferrer" target="_blank"> which weighed in at roughly 103 million tonnes in 2024</a>.</p><h3></h3><br/><p>The plants, which will generate a combined 2.26 GW, will use combined-cycle gas turbines that recapture waste heat from exhaust.<a href="https://www.ge.com/news/press-releases/ha-technology-now-available-industry-first-64-percent-efficiency" target="_blank"> This boosts thermal efficiency to 60 percent and beyond,</a> meaning more fuel is converted to useful energy. Simple-cycle turbines, by contrast, vent the exhaust, which lowers efficiency to around 40 percent.</p><p>Even so, total life-cycle emissions for the Hyperion plants could range from 4 million to over 10 million tonnes of CO2 each year, depending on how frequently the plants are put in use and the final efficiency benchmarks once built. On the high end, that’s as much CO2 as produced by over 2 million passenger cars. Fortunately, not all of Meta’s data centers take the same approach to power. The company has announced a plan to power Prometheus, a large data-center project in Ohio scheduled to come online before the end of 2026, <a href="https://about.fb.com/news/2026/01/meta-nuclear-energy-projects-power-american-ai-leadership/" target="_blank">with nuclear energy</a>.</p><p>But other big tech companies, spurred by the need to build data centers quickly, are taking a less efficient approach.</p><p>xAI’s Colossus 2, located in Memphis, is the most extreme example. <a href="https://www.climateandcapitalmedia.com/35-gas-turbines-no-permits-elon-musks-dirty-xai-secret/" rel="noopener noreferrer" target="_blank">The company trucked dozens of temporary gas-turbine generators to power the site</a> located in a suburban neighborhood. OpenAI, meanwhile, has gas turbines capable of generating up to 300 megawatts <a href="https://www.timesrecordnews.com/story/news/2025/10/14/water-electricity-concerns-addressed-by-stargate-data-center-leaders-in-abilene-texas/86585222007/" rel="noopener noreferrer" target="_blank">at its new Stargate data center in Abilene, Texas</a>, slated to open later in 2026. Both use simple-cycle turbines with a much lower efficiency rating than the combined-cycle plants Entergy will build to power Hyperion.</p><p>Demand for gas turbines is so intense, in fact, that <a href="https://www.spglobal.com/commodity-insights/en/news-research/latest-news/electric-power/052025-us-gas-fired-turbine-wait-times-as-much-as-seven-years-costs-up-sharply" rel="noopener noreferrer" target="_blank">wait times for new turbines are up to seven years</a>. Some data centers <a href="https://spectrum.ieee.org/ai-data-centers" target="_self">are turning toward refurbished jet engines</a> to obtain the turbines they need.</p><h2>AI Racks Tip the Scales</h2><p>The demand for new, reliable power is driven by the power-hungry GPUs inside modern AI data centers.</p><p>In January of 2025, Mark Zuckerberg announced in a post on Facebook that Meta planned to end 2025 <a href="https://techcrunch.com/2025/01/24/mark-zuckerberg-says-meta-will-have-1-3m-gpus-for-ai-by-year-end/" rel="noopener noreferrer" target="_blank">with at least 1.3 million GPUs in service</a>. OpenAI’s Stargate data center <a href="https://www.datacenterdynamics.com/en/news/openai-and-oracle-to-deploy-450000-gb200-gpus-at-stargate-abilene-data-center/" rel="noopener noreferrer" target="_blank">plans to use over 450,000 Nvidia GB200 GPUs</a>, and xAI’s Colossus 2, an expansion of Colossus, <a href="https://www.nextbigfuture.com/2025/09/xai-colossus-2-first-gigawatt-ai-training-data-center.html" rel="noopener noreferrer" target="_blank">is built to accommodate over 550,000 GPUs</a>.</p><p>GPUs, which remain by far the most popular for AI workloads, are bundled into human-scale monoliths of steel and silicon which, much like the data centers built to house them, are rapidly growing in weight, complexity, and power consumption.</p><h3>Memory</h3><br/><img alt="Outlined head with a microchip brain on blue background, symbolizing AI and technology." class="rm-shortcode" data-rm-shortcode-id="7cd8d3faff2d24fa591295b9efd9b1ba" data-rm-shortcode-name="rebelmouse-image" id="70372" loading="lazy" src="https://spectrum.ieee.org/media-library/outlined-head-with-a-microchip-brain-on-blue-background-symbolizing-ai-and-technology.png?id=65350865&width=980"/><p>In addition to raw compute performance, Nvidia GB200 NVL72 racks also require huge amounts of memory. An Nvidia GB200 NVL72 rack may include up to 13.4 terabytes of high-bandwidth memory, which implies a data-center campus at Hyperion’s scale will require at least several dozen petabytes.</p><p>The immense demand has sent memory prices soaring:<a href="https://wccftech.com/dram-prices-have-risen-by-a-whopping-172-this-year-alone/" rel="noopener noreferrer" target="_blank"> The price of DRAM, specifically DDR5, has increased 172 percent in 2025</a>.</p><h3>Power</h3><br/><img alt="" class="rm-shortcode" data-rm-shortcode-id="eaf0380400ba03875bf2ee910f35ab5d" data-rm-shortcode-name="rebelmouse-image" id="5bd7d" loading="lazy" src="https://spectrum.ieee.org/media-library/image.png?id=65350873&width=980"/><p>Hyperion is expected to use 5 gigawatts of power across 11 buildings, which works out to just under 500 megawatts per building, assuming each will be similar to its siblings. That’s enough to power roughly 4.2 million U.S. homes.</p><p>Just one Hyperion data center built at the Richland Parish site will consume twice as much power as xAI’s Colossus which, at the time of its completion in the summer of 2024, was among the largest data centers yet built.</p><h3></h3><br/><p>Nvidia’s <a href="https://www.nvidia.com/en-us/data-center/gb200-nvl72/" target="_blank">GB200 NVL72</a>—a rack-scale system—is currently a leading choice for AI data centers. A single GB200 rack contains 72 GPUs, 36 CPUs, and up to 17 terabytes of memory. It measures 2.2 meters tall, <a href="https://aivres.com/wp-content/uploads/KRS8000v3.1.pdf" target="_blank">tips the scales at up to </a>1,553 kilograms, and consumes about 120 kilowatts—as much as around 100 U.S. homes. And this, according to Nvidia, is just the beginning. The company anticipates future racks could <a href="https://www.tomshardware.com/tech-industry/nvidia-to-boost-ai-server-racks-to-megawatt-scale-increasing-power-delivery-by-five-times-or-more" target="_blank">consume up to a megawatt each</a>.</p><p><a href="https://www.linkedin.com/in/viktorpetik/?originalSubdomain=hr" target="_blank">Viktor Petik</a>, senior vice president of infrastructure solutions at<a href="https://www.vertiv.com/en-us/" rel="noopener noreferrer" target="_blank"> Vertiv</a>, says the rapid change in rack-scale AI systems has forced data centers to adapt. “AI racks consume far more power and weigh more than their predecessors,” says Petik. He adds that data centers must supply racks with multiple power feeds, without taking up extra space.</p><p>The new power demands from rack-scale systems have consequences that are reflected in the design of the data center—even its footprint.</p><p>In 2022 Meta broke ground on a new data center at a campus in Temple, Texas. According to <a href="https://semianalysis.com/" rel="noopener noreferrer" target="_blank">SemiAnalysis</a>, which studies AI data centers, construction began with the intent <a href="https://newsletter.semianalysis.com/p/datacenter-anatomy-part-1-electrical" rel="noopener noreferrer" target="_blank">to build the data center in an H-shaped configuration common to other Meta data centers</a>.</p><h3>LAND</h3><br/><img alt="Black location pin icon on orange background." class="rm-shortcode" data-rm-shortcode-id="a2b2e04f07bd0ed3f60e1f86029497af" data-rm-shortcode-name="rebelmouse-image" id="248cd" loading="lazy" src="https://spectrum.ieee.org/media-library/black-location-pin-icon-on-orange-background.png?id=65351137&width=980"/><h3></h3><br/><p>Meta CEO Mark Zuckerberg kicked off the buzz around Hyperion by saying it would cover a large chunk of Manhattan. Many took that to mean Hyperion would be a single building of that size, which isn’t correct. Hyperion will actually be a cluster of data centers—11 are currently planned—with over 370,000 square meters of floor space. That’s a lot smaller even than New York City’s Central Park, which covers 6 percent of Manhattan.</p><p>Meta has room to grow, however. The Richland Parish site spans 14.7 million square meters in total, which is about a quarter the area of Manhattan. And the 370,000 square meters of floor space Hyperion is expected to provide doesn’t include external infrastructure, such as the three new combined-cycle gas power plants Louisiana utility Entergy is building to power the project.</p><h3></h3><br/><img alt="Map with site layout and regional location in Louisiana, showing roads and distances." class="rm-shortcode" data-rm-shortcode-id="b0cc9253de57aefb96d39a9892c95fe5" data-rm-shortcode-name="rebelmouse-image" id="a41a4" loading="lazy" src="https://spectrum.ieee.org/media-library/map-with-site-layout-and-regional-location-in-louisiana-showing-roads-and-distances.png?id=65352088&width=980"/><h3></h3><br/><p><span>Construction was paused midway in December of 2022, however, </span><a href="https://www.datacenterdynamics.com/en/news/exclusive-after-meta-cancels-odense-data-center-expansion-other-projects-are-being-rescoped/" target="_blank">as part of a company-wide review of its data-center infrastructure</a><span>. Meta decided to knock down the structure it had built and start from scratch. The reasons for this decision were never made public, but analysts believe it was due to the old design’s inability to deliver sufficient electricity to new, power-hungry AI racks. Construction resumed in 2023.</span></p><p>Meta’s replacement ditches the H-shaped building for simple, long, rectangular structures, each flanked by rows of gas-turbine generators. While Meta’s plans are subject to change, Hyperion is currently expected to comprise 11 rectangular data centers, each packed with hundreds of thousands of GPUs, spread across the 13.6-square-kilometer Richland Parish campus.</p><h2>Cooling, and Connecting, at Scale</h2><p>Nvidia’s ultradense AI GPU racks are changing data centers not only with their weight, and power draw, but also with their intense cooling and bandwidth requirements.</p><p>Data centers traditionally use air cooling, but that approach has reached its limits. “Air as a cooling medium is inherently inferior,” says<a href="https://cde.nus.edu.sg/me/staff/lee-poh-seng/" target="_blank"> Poh Seng Lee</a>, head of <a href="https://blog.nus.edu.sg/coolestlab/" rel="noopener noreferrer" target="_blank">CoolestLAB</a>, a cooling research group at the National University of Singapore.</p><p>Instead, going forward, GPUs will rely on liquid cooling. However, that adds a new layer of complexity. “It’s all the way to the facilities level,” says Lee. “You need pumps, which we call a coolant distribution unit. The CDU will be connected to racks using an elaborate piping network. And it needs to be designed for redundancy.” On the rack, pipes connect to cold plates mounted atop every GPU; outside the data-center shell, pipes route through evaporation cooling units. Lee says retrofitting an air-cooled data center is possible but expensive.</p><p>The networking used by AI data centers is also changing to cope with new requirements. Traditional data centers were positioned near network hubs for easy access to the global internet. AI data centers, though, are more concerned with networks of GPUs.</p><p>These connections must sustain high bandwidth with impeccable reliability. Mark Bieberich, a vice president at network infrastructure company Ciena, says its latest fiber-optic transceiver technology,<a href="https://www.ciena.com/products/wavelogic/wavelogic-6" rel="noopener noreferrer" target="_blank"> WaveLogic 6</a>, can provide up to 1.6 terabytes per second of bandwidth per wavelength. A single fiber can support 48 wavelengths in total, and Ciena’s largest customers have hundreds of fiber pairs, placing total bandwidth in the thousands of terabits per second.</p><h3></h3><br/><img alt="a piece of land with a big platform in the middle." class="rm-shortcode" data-rm-shortcode-id="fb6adbcb1ff833934363d6f6ce9cf993" data-rm-shortcode-name="rebelmouse-image" id="63272" loading="lazy" src="https://spectrum.ieee.org/media-library/a-piece-of-land-with-a-big-platform-in-the-middle.jpg?id=65343457&width=980"/><p><span>This is a point where the scale of Meta’s Hyperion, and other large AI data centers, can be deceptive. It seems to imply the physical size of a single data center is what matters. But rather than being a single building,</span><a href="https://datacenters.atmeta.com/richland-parish-data-center/" target="_blank"> Hyperion is actually a set of buildings</a><span> connected by high-speed fiber-optics.</span></p><p>“Interconnecting data centers is absolutely essential,” says Bieberich. “You could think about it as one logical AI training facility, but with geographically distributed facilities.” Nvidia has taken to calling this “scale across,” to contrast it with the idea that data centers must “scale up” to larger singular buildings.</p><h2>The Big but Hazy Future</h2><p>The full scale of the challenges that face Hyperion, and other future AI data centers of similar scale, remain hazy. Nvidia has yet to introduce the rack-scale AI GPU systems it will host. How much power will it demand? What type of cooling will it require? How much bandwidth must be provided? These can only be estimated.</p><p>In the absence of details, the gravity of AI data-center design is pulled toward one certainty: It must be big. New data-center designers are rewriting their rule book to handle power, cooling, and network infrastructure at a scale that would’ve seemed ridiculous five years ago.</p><p>This innovation is fueled by big tech’s fat wallet, which shelled out tens of billions of dollars in 2025 alone, leading to<a href="https://hbr.org/2025/10/is-ai-a-boom-or-a-bubble" target="_blank"> questions about whether the spending is sustainable</a>. For the engineers in the trenches of data-center design, though, it’s viewed as an opportunity to make the impossible possible.</p><p> “I tell my engineers, this is peak. We’re being engineers. We’re being asked complicated questions,” says Stantec’s Carter. “We haven’t got to do that in a long time.” <span class="ieee-end-mark"></span></p><p><em>This article appears in the April 2026 print issue.</em></p>]]></description><pubDate>Tue, 24 Mar 2026 15:00:05 +0000</pubDate><guid>https://spectrum.ieee.org/5gw-data-center</guid><category>Ai</category><category>Power</category><category>Construction</category><category>Data-centers</category><category>Type-cover</category><dc:creator>Matthew S. Smith</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/construction-symbols-on-yellow-background.png?id=65356154&amp;width=980"></media:content></item><item><title>Transforming Data Science With NVIDIA RTX PRO 6000 Blackwell Workstation Edition</title><link>https://spectrum.ieee.org/nvidia-rtx-pro-6000-pny</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/computer-setup-with-a-monitor-displaying-forest-graphics-keyboard-mouse-and-a-sleek-cpu-design.png?id=65315285&width=1200&height=800&coordinates=19%2C0%2C19%2C0"/><br/><br/><p><em>This is a sponsored article brought to you by <a href="https://www.pny.com/" target="_blank">PNY Technologies</a>.</em></p>In today’s data-driven world, data scientists face mounting challenges in preparing, scaling, and processing massive datasets. Traditional CPU-based systems are no longer sufficient to meet the demands of modern AI and analytics workflows. <a href="https://www.pny.com/nvidia-rtx-pro-6000-blackwell-ws?iscommercial=true&utm_source=IEEE+Spectrum+Blog&utm_medium=RTX+PRO+6000+body&utm_campaign=Blackwell+Workstation&utm_id=RTX+PRO+6000" rel="noopener noreferrer" target="_blank">NVIDIA RTX PRO<sup>TM</sup> 6000 Blackwell Workstation Edition</a> offers a transformative solution, delivering accelerated computing performance and seamless integration into enterprise environments.<h2>Key Challenges for Data Science</h2><ul><li><strong>Data Preparation: </strong>Data preparation is a complex, time-consuming process that takes most of a data scientist’s time.</li><li><strong>Scaling: </strong>Volume of data is growing at a rapid pace. Data scientists may resort to downsampling datasets to make large datasets more manageable, leading to suboptimal results.</li><li><strong>Hardware: </strong>Demand for accelerated AI hardware for data centers and cloud service providers (CSPs) is exceeding supply. Current desktop computing resources may not be suitable for data science workflows.</li></ul><h2>Benefits of RTX PRO-Powered AI Workstations</h2><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition delivers ultimate acceleration for data science and AI workflows. These powerful and robust workstations enable real-time rendering, rapid prototyping, and seamless collaboration. With support for up to four <a href="https://www.pny.com/nvidia-rtx-pro-6000-blackwell-max-q?iscommercial=true&utm_source=IEEE+Spectrum+Blog&utm_medium=RTX+PRO+6000+Blackwell+Max-Q+body&utm_campaign=Blackwell+Workstation&utm_id=RTX+PRO+6000" rel="noopener noreferrer" target="_blank">NVIDIA RTX PRO 6000 Blackwell Max-Q Workstation Edition</a> GPUs, users can achieve data center-level performance right at their desk, making even the most demanding tasks manageable.</p><p class="shortcode-media shortcode-media-youtube"> <span class="rm-shortcode" data-rm-shortcode-id="61bf7564ac8304e10487689487367c94" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/jwxxgHsU1jA?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span> <small class="image-media media-caption" placeholder="Add Photo Caption...">PNY is redefining professional computing with the ‪@NVIDIA‬ RTX PRO 6000 Blackwell Workstation Edition, the most powerful desktop GPU ever built. Engineered for unmatched compute power, massive memory capacity, and breakthrough performance, this cutting-edge solution delivers a quantum leap forward in workflow efficiency, enabling professionals to tackle the most demanding applications with ease.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">PNY</small></p><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition empowers data scientists to handle massive datasets, perform advanced visualizations, and support multi-user environments without compromise. It’s ideal for organizations scaling up their analytics or running complex models. NVIDIA RTX PRO 6000 Blackwell Workstation Edition is optimized for AI workflows, leveraging the NVIDIA AI software stack, including CUDA-X, and NVIDIA Enterprise software. These platforms enable zero-code-change acceleration for Python-based workflows and support over 100 AI-powered applications, streamlining everything from data preparation to model deployment.</p><p>Finally, NVIDIA RTX PRO 6000 Blackwell Workstation Edition offers significant advantages in security and cost control. By offloading compute from the data center and reducing reliance on cloud resources, organizations can lower expenses and keep sensitive data on-premises for enhanced protection.</p><h2>Accelerate Every Step of Your Workflow</h2><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition is designed to transform the entire data science pipeline, delivering end-to-end acceleration from data preparation to model deployment. With NVIDIA CUDA-X open-source data science cuDF library and other GPU-accelerated libraries, data scientists can process massive datasets at lightning speed, often achieving up to 50X faster performance compared to traditional CPU-based tools. This means tasks like cleaning data, managing missing values, and engineering features can be completed in seconds, not hours, allowing teams to focus on extracting insights and building better models.</p><p class="pull-quote">NVIDIA RTX PRO 6000 Blackwell Workstation Edition is designed to transform the entire data science pipeline, delivering end-to-end acceleration from data preparation to model deployment</p><p>Exploratory data analysis is elevated with advanced analytics and interactive visualizations, powered by NVIDIA CUDA-X and PyData libraries. These tools enable users to create expansive, responsive visualizations that enhance understanding and support critical decision-making. When it comes to model training, GPU-accelerated XGBoost slashes training times from weeks to minutes, enabling rapid iteration and faster time-to-market AI solutions.</p><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition streamlines collaboration and scalability. With NVIDIA AI Workbench, teams can set up projects, develop, and collaborate seamlessly across desktops, cloud platforms, and data centers. The unified software stack ensures compatibility and robustness, while enterprise-grade hardware maximizes uptime and reliability for demanding workflows.</p><p>By integrating these advanced capabilities, NVIDIA RTX PRO 6000 Blackwell Workstation Edition empowers data scientists to overcome bottlenecks, boost productivity, and drive innovation, making them an essential foundation for modern, enterprise-ready AI development.</p><h2>Performance Benchmarks</h2><p>NVIDIA’s cuDF library offers zero-code change acceleration for pandas, delivering up to 50X performance gains. For example, a join operation that takes nearly 5 minutes on CPU completes in just 14 seconds on GPU. Advanced group by operations drop from almost 4 minutes to just 4 seconds.</p><h2>Enterprise-Ready Solutions from PNY</h2><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" rel="float: left;" style="float: left;"> <img alt="Black PNY logo with stylized uppercase letters on a transparent background." class="rm-shortcode" data-rm-shortcode-id="247ffcd9e141f1fc61c5172c5440d97e" data-rm-shortcode-name="rebelmouse-image" id="170af" loading="lazy" src="https://spectrum.ieee.org/media-library/black-pny-logo-with-stylized-uppercase-letters-on-a-transparent-background.png?id=65315393&width=980"/></p><p>Available from leading OEM manufacturers, NVIDIA RTX PRO 6000 Blackwell Workstation Edition Series GPUs are specifically engineered to meet the rigorous demands of enterprise environments. These systems incorporate NVIDIA Connect-X networking, now available at PNY and a comprehensive suite of deployment and support tools, ensuring seamless integration with existing IT infrastructure.</p><p>Designed for scalability, the latest generation of workstations can tackle complex AI development workflows at scale for training, development, or inferencing. Enterprise-grade hardware maximizes uptime and reliability.</p><p><strong>To learn more about NVIDIA RTX PRO™ Blackwell solutions, </strong><strong>visit:</strong> <a href="https://www.pny.com/professional/software-solutions/blackwell-architecture?utm_source=IEEE+Spectrum+Blog&utm_medium=Blackwell+Desktop+GPUs+learn+more&utm_campaign=Blackwell+Workstation&utm_id=RTX+PRO+6000" target="_blank">NVIDIA RTX PRO Blackwell | PNY Pro | pny.com</a> or email <a href="mailto:gopny@pny.com" target="_blank">GOPNY@PNY.COM</a><strong></strong></p>]]></description><pubDate>Mon, 23 Mar 2026 13:00:04 +0000</pubDate><guid>https://spectrum.ieee.org/nvidia-rtx-pro-6000-pny</guid><category>Artificial-intelligence</category><category>Computing</category><category>Data-science</category><category>Gpu-acceleration</category><category>Ai-workstations</category><category>Nvidia</category><dc:creator>PNY Technologies</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/computer-setup-with-a-monitor-displaying-forest-graphics-keyboard-mouse-and-a-sleek-cpu-design.png?id=65315285&amp;width=980"></media:content></item><item><title>Startups Bring Optical Metamaterials to AI Data Centers</title><link>https://spectrum.ieee.org/optical-metamaterials-ai-data-centers</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-hand-holding-a-microchip-between-thumb-and-forefinger.jpg?id=65322426&width=1200&height=800&coordinates=0%2C83%2C0%2C84"/><br/><br/><p><span>Light-warping physics made “invisibility cloaks” a possibility. Now two startups hope to harness the science underlying this advance to boost the bandwidth of data centers and speed artificial intelligence.</span></p><p>Roughly 20 years ago, scientists developed the <a href="https://www.science.org/doi/10.1126/science.1125907" target="_blank">first</a> <a href="https://www.science.org/doi/10.1126/science.1133628" target="_blank"> structures</a> capable of curving light around objects to conceal them. These are composed of optical <a href="https://spectrum.ieee.org/two-photon-lithography-3d-printing" target="_self">metamaterials</a>—materials with structures smaller than the wavelengths they are designed to manipulate, letting them bend light in unexpected ways.</p><p>The problem with optical cloaks? “There’s no market for them,” says Patrick Bowen, cofounder and CEO of photonic computing startup <a href="https://www.neurophos.com/" target="_blank">Neurophos</a> in Austin, Texas. For instance, each optical cloak typically works only on a single color of light instead of on all visible colors as you might want for stealth applications.</p><p>Now companies are devising more practical uses for the science behind cloaks, such as improving the switches that connect computers in data centers for AI and other cloud services. Increasingly, <a href="https://newsletter.semianalysis.com/p/google-apollo-the-3-billion-game" target="_blank">data centers are looking to use optical circuit switches </a>to overcome the bandwidth limits and power consumption of conventional electronic switches and networks that require converting data between light to electrons multiple times.</p><p class="ieee-inbody-related">RELATED:  <a href="https://spectrum.ieee.org/optical-interconnects-imec-silicon-photonics" target="_blank">Semiconductor Industry Closes in on 400 Gb/s Photonics Milestone</a></p><p>However, today’s optical switching technologies have drawbacks of their own. For instance, ones that depend on silicon photonics face problems with energy efficiency, while those that rely on <a href="https://spectrum.ieee.org/self-assembly" target="_self">microelectromechanical systems (MEMS)</a> can prove unreliable, says Sam Heidari, CEO of optical metasurface startup <a href="https://lumotive.com/" rel="noopener noreferrer" target="_blank">Lumotive</a> in Redmond, Wash.</p><p>Instead, <a href="https://www.nature.com/articles/s44287-024-00136-4" rel="noopener noreferrer" target="_blank">Lumotive has developed metamaterials with adjustable properties</a>. Its new microchip, which debuted 19 March, is covered with copper structures built using standard chipmaking techniques. Between these copper features are <a href="https://spectrum.ieee.org/metasurface-displays" target="_self">liquid crystal</a> elements. The structure of these elements are electronically programmable, just like in liquid crystal displays (LCDs), to alter the optical properties of the metamaterial chip.</p><p>The microchip can precisely steer, lens, shape, and split beams of light reflected off its surface. It can perform all the same functions as multiple optical components with no moving parts in a programmable way in real time, according to Lumotive. “Having no moving parts significantly improves reliability,” Heidari says.</p><p>“We had to go through a lot of R&D at the foundries to not only make our devices functional, but also commercially viable in terms of the right cost and right reliability,” Heidari says.</p><p>The company says its new chips are capable of manipulating not only the industry’s standard of 256 by 256 ports, but could scale up to 10,000 by 10,000. “We think this is game-changing for data centers,” Heidari says. Lumotive plans to launch its first optical switches at the end of 2026.</p><h2>Optical Computing With Metamaterials</h2><p>Similarly, Neurophos hopes its technology may be transformative for artificial intelligence. Since AI is proving energy hungry when run on conventional electronics, scientists are exploring <a href="https://spectrum.ieee.org/optical-neural-networks" target="_self">optical computing</a> as a low-power alternative by processing data with light instead of electrons.</p><p>However, optical processors in the works today are typically far too bulky to achieve a compute density competitive with the best modern electronic processors, Bowen says. Neurophos says it can use metamaterials to build optical modulators—the optical equivalent of a transistor—that are 1/10,000th the size of today’s designs using standard chipmaking processes. “It’s entirely CMOS,” Bowen says. “There are no exotic materials in it.”</p><p>When a laser beam encoding data shines on a Neurophos chip, the way in which each metamaterial element is configured alters the reflected beam to encode results from complex AI tasks. “We basically fit a 1,000- by-1,000 array of optical modulators on a tiny 5-by-5-millimeter area on a chip,” Bowen says. “If you wanted to do that with off-the-shelf silicon photonics, your chip would be a square meter in size.”</p><p>All in all, Bowen claims the Neurophos microchip will offer 50 times greater compute density and 50 times greater energy efficiency than Nvidia’s Blackwell-generation GPU. The company says that hyperscalers—the world’s biggest cloud service providers—will evaluate two upcoming proof-of-concept chips this year. Neurophos is targeting its first systems for early 2028, with production ramping mid-2028.</p>]]></description><pubDate>Thu, 19 Mar 2026 19:19:43 +0000</pubDate><guid>https://spectrum.ieee.org/optical-metamaterials-ai-data-centers</guid><category>Artificial-intelligence</category><category>Data-center</category><category>Optical-switch</category><category>Optical-computing</category><category>Metamaterial</category><category>Metamaterials</category><dc:creator>Charles Q. Choi</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-hand-holding-a-microchip-between-thumb-and-forefinger.jpg?id=65322426&amp;width=980"></media:content></item><item><title>ENIAC, the First General-Purpose Digital Computer, Turns 80</title><link>https://spectrum.ieee.org/eniac-80-ieee-milestone</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/wide-view-of-men-and-women-working-on-the-eniac-in-the-1940s-all-four-walls-from-floor-to-ceiling-host-different-pieces-of-t.jpg?id=65315846&width=1200&height=800&coordinates=0%2C83%2C0%2C84"/><br/><br/><p>Happy 80th anniversary, ENIAC! The <a href="https://penntoday.upenn.edu/news/penns-eniac-worlds-first-electronic-computer-turns-80" rel="noopener noreferrer" target="_blank">Electronic Numerical Integrator and Computer</a>, the first large-scale, general-purpose, programmable electronic digital computer, helped shape our world.</p><p>On 15 February 1946, ENIAC—developed in the <a href="https://facilities.upenn.edu/maps/locations/moore-school-building" rel="noopener noreferrer" target="_blank">Moore School of Electrical Engineering</a> at the <a href="https://www.upenn.edu/" rel="noopener noreferrer" target="_blank">University of Pennsylvania</a>, in Philadelphia—was publicly demonstrated for the first time. Although primitive by today’s standards, ENIAC’s purely electronic design and programmability were breakthroughs in computing at the time. ENIAC made high-speed, general-purpose computing practicable and laid the foundation for today’s machines.</p><p>On the eve of its unveiling, the <a href="https://www.war.gov/" rel="noopener noreferrer" target="_blank">U.S. Department of War</a> issued a<a href="https://americanhistory.si.edu/comphist/pr1.pdf" rel="noopener noreferrer" target="_blank"> news release</a> hailing it as a new machine “expected to revolutionize the mathematics of engineering and change many of our industrial design methods.” Without a doubt, electronic computers have transformed engineering and mathematics, as well as practically every other domain, including politics and spirituality.</p><p>ENIAC’s success ushered the modern computing industry and laid the foundation for today’s digital economy. During the past eight decades, computing has grown from a niche scientific endeavor into an engine of economic growth, the backbone of billion-dollar enterprises, and a catalyst for global innovation. Computing has led to a chain of innovations and developments such as stored programs, semiconductor electronics, integrated circuits, networking, software, the Internet, and distributed large-scale systems.</p><h2>Inside the ENIAC</h2><p>The motivation for developing ENIAC was the <a href="https://www.pbs.org/wgbh/aso/databank/entries/dt45en.html" rel="noopener noreferrer" target="_blank">need for faster computation</a> during World War II. The U.S. military wanted to produce extensive artillery firing tables for field gunners to quickly determine settings for a specific weapon, a target, and conditions. Calculating the tables by hand took “<a href="https://cacm.acm.org/blogcacm/computers-were-originally-humans/" rel="noopener noreferrer" target="_blank">human computers</a>” several days, and the available mechanical machines were far too slow to meet the demand.</p><h3>80 Years of Electronic Computer Milestones </h3><br/><h4>1946</h4><p><a href="https://www.britannica.com/technology/ENIAC" rel="noopener noreferrer" target="_blank"><strong>ENIAC operational</strong></a></p><p>Birth of electronic computing</p><h4>1951</h4><p><a href="https://www.britannica.com/technology/UNIVAC" target="_blank"><strong>UNIVAC I</strong></a></p><p><a href="https://www.britannica.com/technology/UNIVAC" target="_blank"></a>Start of commercial computing</p><h4>1958</h4><p><a href="https://www.synopsys.com/glossary/what-is-integrated-circuit.html" target="_blank"><strong>Integrated circuit</strong></a></p><p>Foundation for modern computer hardware</p><h4>1964</h4><p><a href="https://www.ibm.com/history/system-360" rel="noopener noreferrer" target="_blank"><strong>IBM System/360</strong></a></p><p>Popular mainframe computer</p><h4>1970</h4><p><a href="https://en.wikipedia.org/wiki/PDP-11" rel="noopener noreferrer" target="_blank"><strong>Programmed Data Processor (PDP-11)</strong></a></p><p>Popular 16-bit minicomputer</p><h4>1971</h4><p><a href="https://computer.howstuffworks.com/microprocessor.htm" rel="noopener noreferrer" target="_blank"><strong>Intel 4004</strong></a></p><p>Beginning of the microprocessor and microcomputer era</p><h4>1975</h4><p><a href="https://en.wikipedia.org/wiki/Cray-1" rel="noopener noreferrer" target="_blank"><strong>Cray-1</strong></a></p><p>First supercomputer</p><h4>1977</h4><p><a href="https://www.stromasys.com/resources/vax-computer-systems-an-in-depth-guide/" rel="noopener noreferrer" target="_blank"><strong>VAX</strong></a></p><p>Popular 32-bit minicomputer</p><h4>1981</h4><p><a href="https://en.wikipedia.org/wiki/IBM_Personal_Computer" rel="noopener noreferrer" target="_blank"><strong>IBM PC</strong></a></p><p>Personal and small-business computing</p><h4>1989</h4><p><a href="https://home.cern/science/computing/birth-web" rel="noopener noreferrer" target="_blank"><strong>World Wide Web</strong></a></p><p>Digital communication, interaction, and transaction (e-commerce)</p><h4>2002</h4><p><a href="https://en.wikipedia.org/wiki/Amazon_Web_Services" rel="noopener noreferrer" target="_blank"><strong>Amazon Web Services</strong></a></p><p>Beginning of the cloud computing revolution</p><h4>2010</h4><p><a href="https://en.wikipedia.org/wiki/IPad" rel="noopener noreferrer" target="_blank"><strong>Apple iPad</strong></a></p><p>Handheld computer/tablet</p><h4>2010</h4><p><a href="https://www.ibm.com/think/topics/industry-4-0" rel="noopener noreferrer" target="_blank"><strong>Industry 4.0</strong></a></p><p>Delivered real-time decision-making, smart manufacturing, and logistics</p><h4>2016</h4><p><a href="https://www.livescience.com/55642-reprogrammable-quantum-computer-created.html" rel="noopener noreferrer" target="_blank"><strong>First reprogrammable quantum computer demonstrated</strong></a></p><p>Ignited interest in quantum computing</p><h4>2023</h4><p><a href="https://en.wikipedia.org/wiki/Generative_artificial_intelligence" rel="noopener noreferrer" target="_blank"><strong>Generative AI boom</strong></a></p><p>Widespread use of GenAI by individuals, businesses, and academia</p><h4>2026</h4><p><a href="https://penntoday.upenn.edu/news/penns-eniac-worlds-first-electronic-computer-turns-80" rel="noopener noreferrer" target="_blank"><strong>ENIAC’s 80th anniversary</strong></a></p><p>80 years of computing evolution</p><h3></h3><br/><p>In 1942 <a href="https://www.britannica.com/biography/John-Mauchly" target="_blank">John Mauchly</a>, an associate professor of electrical engineering at Penn’s Moore School, suggested using vacuum tubes to speed up computer calculations. Following up on his theory, the U.S. Army <a href="https://en.wikipedia.org/wiki/Ballistic_Research_Laboratory" target="_blank">Ballistic Research Laboratory</a>, which was responsible for providing artillery settings to soldiers in the field, commissioned Mauchly and his colleagues<a href="https://www.britannica.com/biography/J-Presper-Eckert-Jr" target="_blank"> </a><a href="https://ethw.org/J._Presper_Eckert" rel="noopener noreferrer" target="_blank">J. Presper Eckert</a> and <a href="https://ethw.org/Adele_Katz_Goldstine" target="_blank">Adele Katz Goldstine</a>, to work on a new high-speed computer. Eckert was a lab instructor at Moore, and Goldstine became one of ENIAC’s programmers. It took them a year to design ENIAC and 18 months to build it.</p><p>The computer contained about 18,000 vacuum tubes, which were cooled by 80 air blowers. More than 30 meters long, it filled a 9 m by 15 m room and weighed about 30 kilograms. It consumed as much electricity as a small town.</p><p>Programming the machine was <a href="https://www.pbs.org/wgbh/aso/databank/entries/dt45en.html" target="_blank">difficult</a>. ENIAC did not have stored programs, so to reprogram the machine, operators manually reconfigured cables with switches and plugboards, a process that took several days.</p><p>By the 1950s, large universities either had acquired or built their own machines to rival ENIAC. The schools included <a href="https://www.cam.ac.uk/" rel="noopener noreferrer" target="_blank">Cambridge</a> (EDSAC), <a href="https://www.mit.edu/" rel="noopener noreferrer" target="_blank">MIT</a> (Whirlwind), and <a href="https://www.princeton.edu/" rel="noopener noreferrer" target="_blank">Princeton</a> (IAS). Researchers used the computers to model physical phenomena, solve mathematical problems, and perform simulations.</p><p>After almost nine years of operation, ENIAC officially was decommissioned on 2 October 1955.</p><p><a href="https://mitpress.mit.edu/9780262535175/eniac-in-action/" rel="noopener noreferrer" target="_blank"><em>ENIAC in Action: Making and Remaking the Modern Computer</em></a>, a book by <a href="https://uwm.edu/history/about/directory/haigh-thomas/" rel="noopener noreferrer" target="_blank">Thomas Haigh</a>, <a href="https://mitpress.mit.edu/author/mark-priestley-15374/" rel="noopener noreferrer" target="_blank">Mark Priestley</a>, and <a href="https://www.researchgate.net/scientific-contributions/Crispin-Rope-2045495041" rel="noopener noreferrer" target="_blank">Crispin Rope</a>,<em> </em>describes the design, construction, and testing processes and dives into its afterlife use. The book also outlines the complex relationship between ENIAC and its designers, as well as the revolutionary approaches to computer architecture.</p><p>In the early 1970s, there was a controversy over who invented the electronic computer and who would be assigned the patent. In 1973 <a href="https://en.wikipedia.org/wiki/Earl_R._Larson" rel="noopener noreferrer" target="_blank">Judge Earl Richard Larson</a> of U.S. District Court in Minnesota ruled in the <a href="https://en.wikipedia.org/wiki/Honeywell,_Inc._v._Sperry_Rand_Corp." rel="noopener noreferrer" target="_blank">Honeywell <em><em>v.</em></em> Sperry Rand</a> case that Eckert and Mauchly did not invent the automatic electronic digital computer but instead had derived their subject matter from a <a href="https://jva.cs.iastate.edu/operation.php" rel="noopener noreferrer" target="_blank">computer</a> prototyped in 1939 by <a href="https://history-computer.com/people/john-vincent-atanasoff-complete-biography/" rel="noopener noreferrer" target="_blank">John Vincent Atanasoff</a> and Clifford Berry at Iowa State College (now <a href="https://www.iastate.edu/" rel="noopener noreferrer" target="_blank">Iowa State University</a>). The ruling granted Atanasoff legal recognition as the inventor of the first electronic digital computer.</p><h2>IEEE’s ENIAC Milestone</h2><p>In 1987 IEEE<a href="https://ethw.org/Milestones:Electronic_Numerical_Integrator_and_Computer,_1946" rel="noopener noreferrer" target="_blank"> designated ENIAC</a> as an IEEE Milestone, citing it as “a major advance in the history of computing” and saying the machine “established the practicality of large-scale electronic digital computers and strongly influenced the development of the modern, stored-program, general-purpose computer.”</p><p>The commemorative Milestone plaque is displayed at the Moore School, by the entrance to the classroom where ENIAC was built.</p><h3></h3><br/><p>“The ENIAC legacy heralded the computer age, transforming not only science and industry but also education, research, and human communication and interaction.”</p><h3></h3><br/><p><br/></p><p>A <a href="https://ieeexplore.ieee.org/document/476557" rel="noopener noreferrer" target="_blank">paper on the machine</a>, published in 1996 in <a href="https://ieeexplore.ieee.org/document/476557" rel="noopener noreferrer" target="_blank"><em>IEEE Annals of the History of Computing</em></a> and available in the <a href="https://ieeexplore.ieee.org/document/6461145" rel="noopener noreferrer" target="_blank">IEEE Xplore Digital Library</a>, is a valuable source of technical information.</p><p>“<a href="https://www.computer.org/csdl/magazine/an/2006/02/man2006020004/13rRUB6Sq2p" rel="noopener noreferrer" target="_blank">The Second Life of ENIAC</a><em>,”</em> an article published in the annals in 2006, covers a lesser-known chapter in the machine’s history, about how it evolved from a static system—configured and reconfigured through laborious cable plugging—into a precursor of today’s stored-program computers.</p><p>A classic <a href="https://www2.seas.gwu.edu/~mfeldman/csci1030/summer08/eniac2.pdf" rel="noopener noreferrer" target="_blank">history paper on ENIAC</a> was published in the December 1995 <a href="https://technologyandsociety.org/" rel="noopener noreferrer" target="_blank"><em>IEEE Technology and Society Magazine</em></a>.</p><p>The IEEE <a href="https://spectrum.ieee.org/ebooks/ieee-anniversary-book/" target="_self"><em>Inspiring Technology: 34 Breakthroughs</em></a> book, published in 2023, features an ENIAC chapter.</p><h2>The women behind ENIAC</h2><p>One of the most remarkable aspects of the ENIAC story is the pivotal role women played, according to the book <a href="https://www.amazon.com/Proving-Ground-Untold-Programmed-Computer/dp/1538718286" rel="noopener noreferrer" target="_blank"><em>Proving Ground: The Untold Story of the Six Women Who Programmed the World’s First Modern Computer</em></a><em>, </em>highlighted in an <a href="https://spectrum.ieee.org/the-women-behind-eniac" target="_self">article</a> in <a href="https://spectrum.ieee.org/the-institute/" target="_self"><em>The Institute</em></a>. There were no “programmers” at that time; only schematics existed for the computer. Six women, known as the ENIAC 6, became the machine’s first programmers.</p><p>The ENIAC 6 were <a href="https://en.wikipedia.org/wiki/Kathleen_Antonelli" rel="noopener noreferrer" target="_blank">Kathleen Antonelli</a>, <a href="https://en.wikipedia.org/wiki/Jean_Bartik" rel="noopener noreferrer" target="_blank">Jean Bartik</a>, <a href="https://ethw.org/Betty_Holberton" rel="noopener noreferrer" target="_blank">Betty Holberton</a>, <a href="https://ethw.org/Marlyn_Meltzer" rel="noopener noreferrer" target="_blank">Marlyn Meltzer</a>, <a href="https://ethw.org/Frances_Spence" rel="noopener noreferrer" target="_blank">Frances Spence</a>, and <a href="https://ethw.org/Ruth_Teitelbaum" rel="noopener noreferrer" target="_blank">Ruth Teitelbaum</a>.</p><p>“These six women found out what it took to run this computer, and they really did incredible things,” a Penn professor, <a href="https://www.cis.upenn.edu/~mitch/" rel="noopener noreferrer" target="_blank">Mitch Marcus</a>, said in a <a href="https://www.phillyvoice.com/70-years-ago-six-philly-women-eniac-digital-computer-programmers/" rel="noopener noreferrer" target="_blank">2006 PhillyVoice article</a>. Marcus teaches in Penn’s computer and information science department.</p><p>In 1997 all six female programmers were<a href="https://www.witi.com/halloffame/298369/ENIAC-Programmers-Kathleen---/" rel="noopener noreferrer" target="_blank"> inducted</a> into the <a href="https://www.witi.com/halloffame/" rel="noopener noreferrer" target="_blank">Women in Technology International Hall of Fame</a>, in Los Angeles.</p><p>Two other women contributed to the programming. Goldstine wrote ENIAC’s five-volume manual, and <a href="https://en.wikipedia.org/wiki/Kl%C3%A1ra_D%C3%A1n_von_Neumann" rel="noopener noreferrer" target="_blank">Klára Dán von Neumann</a>, wife of <a href="https://ethw.org/John_von_Neumann" rel="noopener noreferrer" target="_blank">John von Neumann</a>, helped train the programmers and debug and verify their code.</p><p>To honor the<a href="https://www.computer.org/volunteering/awards/pioneer/about-women-of-eniac" rel="noopener noreferrer" target="_blank"> women of ENIAC</a>, the <a href="https://www.computer.org/" rel="noopener noreferrer" target="_blank">IEEE Computer Society</a> established the annual<a href="https://www.computer.org/volunteering/awards/pioneer" rel="noopener noreferrer" target="_blank"> Computer Pioneer Award</a> in 1981. Eckert and Mauchly were among the award’s first recipients. In 2008 Bartik was honored with the award. Nominations are open to all professionals, regardless of gender.</p><h2>An ENIAC replica</h2><p>Last year a group of 80 autistic students, ages 12 to 16, from<a href="https://www.psacademyarizona.com/" rel="noopener noreferrer" target="_blank"> PS Academy Arizona</a>, in Gilbert, <a href="https://www.msn.com/en-us/news/technology/how-80-autistic-students-built-an-amazing-replica-of-the-ginormous-eniac-computer/ar-AA1UMKKE" rel="noopener noreferrer" target="_blank">recreated the ENIAC</a> using 22,000 custom parts. It took the students almost six months to assemble.</p><p>A ceremony was held in January to display their creation. The full-scale <a href="https://www.theregister.com/2026/01/21/eniac_model_build/" rel="noopener noreferrer" target="_blank">replica features</a> actual-size panels made from layered cardboard and wood. Although all electronic components are simulated, they are not electrically active. The machine, illuminated by hundreds of LEDs, is accompanied by a soundtrack that simulates the deep hum of ENIAC’s transformers and the rhythmic clicking of relays.</p><p><strong></strong></p><h3></h3><br><img alt="A white woman using a computer-adding machine in the 1940\u2019s. The device resembles a bulky typewriter and prints large stacks of paper with tabulated answers." class="rm-shortcode" data-rm-shortcode-id="fea0fb9da93e75542fd5b85964251c33" data-rm-shortcode-name="rebelmouse-image" id="36a08" loading="lazy" src="https://spectrum.ieee.org/media-library/a-white-woman-using-a-computer-adding-machine-in-the-1940-u2019s-the-device-resembles-a-bulky-typewriter-and-prints-large-stack.jpg?id=65315890&width=980"/><h3></h3><br/><p>“Every major unit, accumulators, function tables, initiator, and master programmer is present and placed exactly where it was on the original machine,” Tom Burick, the teacher who mentored the project, said at the ceremony.</p><p>The replica, still on display at the school, is expected to be moved to a more permanent spot in the near future.</p><h2>ENIAC’s legacy</h2><p>ENIAC’s significance is both technical and symbolic. Technically, it marks the beginning of the chain of innovations that created today’s computational infrastructure. Symbolically, it made governments, militaries, universities, and industry view computation as a tool for improvement and for innovative applications that had previously been impossible. It marked a tectonic shift in the way humans approach problem-solving, modeling, and scientific reasoning.</p><p>The ENIAC legacy heralded the computer age, transforming not only science and industry but also education, research, and human communication and interaction.</p><p>As Eckert is reported to have said, “There are two epochs in computer history: Before ENIAC and After ENIAC.”</p><h2>Coevolution of programming languages</h2><p>The remarkable evolution of computer hardware during the past 80 years has been sparked by advances in programming languages—the essential drivers of computing.</p><p>From the manual rewiring of ENIAC to the orchestration of intelligent, distributed systems, programming languages have steadily evolved to make computers more powerful, expressive, and accessible.</p><h3>Lessons From Computing’s Remarkable Journey</h3><br/><p>Computing history teaches us that flexibility, accessibility, collaboration, sound governance, and forward thinking are essential for sustained technological progress. In a <a href="https://cacm.acm.org/blogcacm/what-past-computing-breakthroughs-teach-us-about-ai/" target="_blank">recent <em><em>Communications of the ACM</em></em> article</a>, <a href="https://www.linkedin.com/in/richa28gupta/" target="_blank">Richa Gupta</a> identified four historic shifts that led to computing’s rapid, transformative progress:</p><ol><li>Programmable machines taught us that flexibility is key; technologies that adapt and are repurposed scale better.</li><li>The Internet showed that connection and standard protocols drive explosive growth but also bring new risks such as data security issues, invasion of privacy, and misuse.</li><li>Personal computers illustrated that accessibility and usability matter more than raw power. When nonexperts can use a tool easily, adoption rises.</li><li>The open-source movement revealed that collaborative innovation accelerates growth and helps spot problems early.</li></ol></br><h2>Predictions for computing in the decades ahead</h2><p>The evolution of computing will continue along multiple trajectories, with the emphasis moving from generalization to specialization (for AI, graphics, security, and networking), from monolithic system design to modular integration, and from performance-centric metrics alone to energy efficiency and sustainability as primary objectives.</p><p>Increasingly, security will be built into hardware by design. Computing paradigms will expand beyond traditional deterministic models to embrace probabilistic, approximate, and hybrid approaches for certain tasks.</p><p>Those developments will usher in a new era of computing and a new class of applications.</p>]]></description><pubDate>Wed, 18 Mar 2026 18:00:05 +0000</pubDate><guid>https://spectrum.ieee.org/eniac-80-ieee-milestone</guid><category>Ieee-history</category><category>Eniac</category><category>Computing</category><category>Computers</category><category>History-of-technology</category><category>Type-ti</category><dc:creator>San Murugesan</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/wide-view-of-men-and-women-working-on-the-eniac-in-the-1940s-all-four-walls-from-floor-to-ceiling-host-different-pieces-of-t.jpg?id=65315846&amp;width=980"></media:content></item><item><title>Wanted: Europe’s Missing Cloud Provider</title><link>https://spectrum.ieee.org/europe-cloud-sovereignty</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/abstract-pixelation-of-the-european-union-s-flag.jpg?id=65298877&width=1200&height=800&coordinates=156%2C0%2C156%2C0"/><br/><br/><p>Looming over the <a href="https://spectrum.ieee.org/free-space-optical-link-taara" target="_self">internet lasers</a> and <a href="https://www.pcmag.com/news/hands-on-with-oukitel-wp63-mwc-2026" rel="noopener noreferrer" target="_blank">firestarting phones</a> companies were touting at Mobile World Congress in Barcelona this month, was a more nebulous but much larger announcement: a pan-European cloud called <a href="https://www.euronews.com/next/2026/03/03/europe-unites-to-build-sovereign-cloud-and-ai-infrastructure-to-stop-reliance-on-us" rel="noopener noreferrer" target="_blank">EURO-3C</a>.</p><p>EURO-3C’s backers—Spanish telecoms giant Telefónica, dozens of other European companies, and the European Commission (EC)—aim to fill a gap. U.S.-based cloud giants dominate in the EU, and European policymakers want their growing portfolio of digital government services on a “sovereign cloud” under full EU control.</p><p>But the EU lacks a real equivalent to the likes of AWS or Microsoft Azure. Indeed, any effort to build one will inevitably run up against the same U.S. cloud giants.</p><p>Just four U.S.-based hyperscalers—AWS, Microsoft Azure, Google Cloud, and IBM Cloud—together account for<a href="https://www.ceps.eu/disk-backup-to-the-cloud-is-a-gaping-vulnerability-in-the-eus-security/" rel="noopener noreferrer" target="_blank"> some 70 percent of EU cloud services</a>. This is despite the fact that the 2018 U.S. <a href="https://en.wikipedia.org/wiki/CLOUD_Act" rel="noopener noreferrer" target="_blank">CLOUD Act</a> allows U.S. federal law enforcement—at least in theory—to compel U.S.-based firms to hand over data that’s stored abroad. </p><h2>Who Do You Trust?</h2><p>But those hypothetical risks to digital services have become more real as transatlantic relations have soured under the second Trump administration. The U.S. has <a href="https://www.cbc.ca/news/politics/greenland-us-trump-canada-governor-general-mary-simon-9.7119074" rel="noopener noreferrer" target="_blank">openly threatened</a> to invade an EU member state and <a href="https://euobserver.com/19745/eu-rejects-us-claims-of-censorship-over-tech-rules-after-visa-bans/" rel="noopener noreferrer" target="_blank">sanctioned</a> a European Commissioner for passing legislation the White House dislikes. </p><p>After the White House sanctioned the Netherlands-based International Criminal Court in February 2025, Court staffers <a href="https://apnews.com/article/icc-trump-sanctions-karim-khan-court-a4b4c02751ab84c09718b1b95cbd5db3" rel="noopener noreferrer" target="_blank">claimed</a> Microsoft locked the Court’s chief prosecutor out of his email (Microsoft<a href="https://www.politico.eu/article/microsoft-did-not-cut-services-international-criminal-court-president-american-sanctions-trump-tech-icc-amazon-google/" rel="noopener noreferrer" target="_blank"> has denied this</a>). Around the same time, the U.S. <a href="https://kyivindependent.com/us-threatens-to-shut-off-starlink-if-ukraine-wont-sign-minerals-deal-sources-tell-reuters/" rel="noopener noreferrer" target="_blank">reportedly threatened</a> to sever EU ally Ukraine’s access to crucial Starlink satellite internet as leverage during trade negotiations.</p><p>“The geopolitical risk isn’t just the most extreme form of a doomsday ‘kill switch’ where Washington turns off Europe’s internet,” says <a href="https://fermigier.com/" rel="noopener noreferrer" target="_blank">Stéfane Fermigier</a> of <a href="https://euro-stack.com/pages/about" rel="noopener noreferrer" target="_blank">EuroStack</a>, an industry group that supports European digital independence. “It is the selective degradation of services and a total lack of retaliatory leverage.”</p><p>What, then, is the EU to do? <a href="https://blog.datacenter-paris.com/2026/01/24/liste-des-datacenters-secnumcloud-en-france-hebergement-souverain-pour-donnees-sensibles/" rel="noopener noreferrer" target="_blank">France</a> offers an example. Even before 2025, France implemented <a href="https://www.spscommerce.com/eur/blog/what-is-secnumcloud-and-does-my-company-need-to-qualify/" rel="noopener noreferrer" target="_blank">harsh restrictions</a> on non-EU cloud providers in public services—providers must locate data in the EU, rely on EU-based staff, and may not have majority non-EU shareholders. Now, EU policymakers are following France’s lead.</p><p>In October 2025, the EC issued a two-part <a href="https://commission.europa.eu/document/09579818-64a6-4dd5-9577-446ab6219113_en" rel="noopener noreferrer" target="_blank">framework</a> for judging cloud providers bidding for public-sector contracts. In the first part, the framework lays out a sort of sovereignty ladder. The more that a provider is subject to EU law, the higher its sovereignty level on this ladder. Any prospective bidder must first meet a certain level, depending on the tender.</p><p>Qualifying bidders then move to the second part, where their “sovereignty” is scored in more detail. Using too much proprietary software; over-relying on supply chains from outside the EU; having non-EU support staff; liability to non-EU laws like the CLOUD Act: All hurt a bidder’s score. </p><p>The framework was created for <a href="https://commission.europa.eu/news-and-media/news/commission-moves-forward-cloud-sovereignty-eur-180-million-tender-2025-10-10_en" rel="noopener noreferrer" target="_blank">one tender</a>, but observers say it sets a major precedent. Cloud providers bidding for state contracts across Europe may need to follow it, and it may influence legislation on both national and EU-wide levels.</p><h2>A Question of Scale</h2><p>Who, then, will receive high marks? At the moment, the answer is not simple. The EU cloud scene is quite fragmented. Numerous modest EU providers offer “sovereign cloud” services—such as Deutsche Telekom’s T-Systems, OVHcloud, and Scaleway—but <a href="https://onlinelibrary.wiley.com/doi/full/10.1002/poi3.358" rel="noopener noreferrer" target="_blank">none are on the scale</a> of AWS or Google Cloud.</p><p>Inertia is on the side of the U.S. cloud giants, which can invest in their infrastructure and services on a far grander scale than their European counterparts. Some U.S. providers <a href="https://aws.amazon.com/blogs/security/aws-european-sovereign-cloud-achieves-first-compliance-milestone-soc-2-and-c5-reports-plus-seven-iso-certifications/" rel="noopener noreferrer" target="_blank">now offer</a> cloud services they say comply with the Commission’s “cloud sovereignty” demands.</p><p>Some European observers, like EuroStack, <a href="https://euro-stack.com/blog/2025/10/cloud-sovereignty-framework-comparison" rel="noopener noreferrer" target="_blank">say</a> such promises are hollow so long as a provider’s parent company is subject to the likes of the CLOUD Act and loopholes in the Commission’s process remain open. An AWS spokesperson told <em>IEEE </em><em>Spectrum</em> it had not disclosed any non-US enterprise or government data to the U.S. government under the CLOUD Act; a Google spokesperson said that its most sensitive EU offerings “are subject to local laws, not U.S. law”.</p><p>Even if a project like EURO-3C can offer a large-scale alternative, the U.S. cloud giants have another sort of inertia. Many developers—and many public purchasers of their services—will need convincing to leave behind a familiar environment.</p><p>“If you look at AWS, you look at Google, they’ve created some super technology. It’s very convenient, it’s easy to use,” says <a href="https://nl.linkedin.com/in/arnoldjuffer" rel="noopener noreferrer" target="_blank">Arnold Juffer</a>, CEO of the Netherlands-based cloud provider <a href="https://nebul.com/" rel="noopener noreferrer" target="_blank">Nebul</a>. “Once you’re in that platform, in that ecosystem, it’s very hard to get out.”</p><p><a href="https://bisi.org.uk/martyna-chmura" rel="noopener noreferrer" target="_blank">Martyna Chmura</a>, an analyst at the Bloomsbury Intelligence and Security Institute, a London-based think tank, sees some EU developers taking a mixed approach. “Many organizations are already moving toward multicloud setups, using European or sovereign providers for sensitive workloads while still relying on hyperscalers for certain services,” she says.</p><p>In that case, the EU’s top-down demands may encourage developers to use EU providers for sensitive applications—like government services, transport, autonomous vehicles, and some industrial automation—even if it’s inconvenient in the short term, or if it causes even more fragmentation of the EU cloud scene. “Running systems across different platforms can increase integration costs and make security and data governance more complicated. In some cases, organisations could lose some of the efficiency and cost advantages that come from using large hyperscale platforms,” Chmura says.</p><p>“Overall, the EU appears willing to accept some of these trade-offs,” Chmura says.</p>]]></description><pubDate>Tue, 17 Mar 2026 11:00:06 +0000</pubDate><guid>https://spectrum.ieee.org/europe-cloud-sovereignty</guid><category>Cloud-computing</category><category>Data-security</category><category>Data-privacy</category><dc:creator>Rahul Rao</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/abstract-pixelation-of-the-european-union-s-flag.jpg?id=65298877&amp;width=980"></media:content></item><item><title>With Nvidia Groq 3, the Era of AI Inference Is (Probably) Here</title><link>https://spectrum.ieee.org/nvidia-groq-3</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-man-in-all-black-presents-in-front-of-a-large-screen-which-compares-a-large-rectangular-chip-labelled-rubin-gpu-with-a-square.jpg?id=65298681&width=1200&height=800&coordinates=156%2C0%2C156%2C0"/><br/><br/><p>This week, over 30,000 people are descending upon San Jose, Calif., to attend<a href="https://www.nvidia.com/gtc/" rel="noopener noreferrer" target="_blank">Nvidia GTC</a>, the so-called Superbowl of AI—a nickname that may or may not have been coined by Nvidia. At the main event Jensen Huang, Nvidia CEO, took the stage to announce (among other things) a new line of<a href="https://spectrum.ieee.org/nvidia-rubin-networking" target="_self">next-generation Vera Rubin</a> chips that represent a first for the GPU giant: a chip designed specifically to handle AI inference. The Nvidia Groq 3 language processing unit (LPU) incorporates intellectual property Nvidia<a href="https://groq.com/newsroom/groq-and-nvidia-enter-non-exclusive-inference-technology-licensing-agreement-to-accelerate-ai-inference-at-global-scale" rel="noopener noreferrer" target="_blank">licensed</a> from the startup Groq last Christmas Eve for US $20 billion.</p><p>“Finally, AI is able to do productive work, and therefore the inflection point of inference has arrived,” Huang told the crowd. “AI now has to think. In order to think, it has to inference. AI now has to do; in order to do, it has to inference.”</p><p>Training and inference tasks have distinct computational requirements. While training can be done on huge amounts of data at the same time and can take weeks, inference must be run on a user’s query when it comes in. Unlike training, inference doesn’t require running costly<a href="https://spectrum.ieee.org/what-is-deep-learning/backpropagation" target="_self">backpropagation</a>. With inference, the most important thing is low latency—users expect the chatbot to answer quickly, and for thinking or reasoning models, inference runs many times before the user even sees an output.</p><p>Over the past few years, inference-specific chip startups were experiencing a sort of Cambrian explosion, with different companies exploring distinct approaches to speed up the task. The startups include<a href="https://www.d-matrix.ai/" rel="noopener noreferrer" target="_blank">D-matrix</a>, with digital in-memory compute;<a href="https://www.etched.com/" rel="noopener noreferrer" target="_blank">Etched</a>, with an ASIC for transformer inference;<a href="https://rain.ai/" rel="noopener noreferrer" target="_blank">RainAI</a>, with neuromorphic chips;<a href="https://en100.enchargeai.com/" rel="noopener noreferrer" target="_blank">EnCharge</a>, with analog in-memory compute;<a href="https://www.tensordyne.ai/" rel="noopener noreferrer" target="_blank">Tensordyne</a>, with logarithmic math to make AI computations more efficient;<a href="https://furiosa.ai/" rel="noopener noreferrer" target="_blank">FuriosaAI</a>, with hardware optimized for tensor operation rather than vector-matrix multiplication, and others.</p><p>Late last year, it looked like Nvidia had picked one of the winners among the crop of inference chips when it announced its deal with Groq. The Nvidia Groq 3 LPU reveal came a mere two and a half months after, highlighting the urgency of the growing inference market.</p><h2>Memory bandwidth and data flow</h2><p>Groq’s approach to accelerating inference relies on interleaving processing units with memory units on the chip. Instead of relying on high-bandwidth memory (HBM) situated next to GPUs, it leans on SRAM memory integrated within the processor itself. This design greatly simplifies the flow of data through the chip, allowing it to proceed in a streamlined, linear fashion.</p><p>“The data actually flows directly through the SRAM,”<a href="https://www.linkedin.com/in/markheaps/" rel="noopener noreferrer" target="_blank">Mark Heaps</a> said at the Supercomputing conference in 2024. Heaps was a chief technology evangelist at Groq at the time and is now director of developer marketing at Nvidia. “When you look at a multicore GPU, a lot of the instruction commands need to be sent off the chip, to get into memory and then come back in. We don’t have that. It all passes through in a linear order.”</p><p>Using SRAM allows that linear data flow to happen exceptionally fast, leading to the low latency required for inference applications. “The LPU is optimized strictly for that extreme low latency token generation,” says<a href="https://www.linkedin.com/in/ian-buck-19201315/" rel="noopener noreferrer" target="_blank">Ian Buck</a>, VP and general manager of hyperscale and high-performance computing at Nvidia.</p><p>Comparing the Rubin GPU and Groq 3 LPU side by side highlights the difference. The Rubin GPU has access to a whopping 288 gigabytes of HBM and is capable of 50 quadrillion floating-point operations per second (petaFLOPS) of 4-bit computation. The Groq 3 LPU contains a mere 500 megabytes of SRAM memory and is capable of 1.2 petaFLOPS of 8-bit computation. On the other hand, while the Rubin GPU has a memory bandwidth of 22 terabytes per second, at 150 TB/s the Groq 3 LPU is seven times as fast. The lean, speed-focused design is what allows the LPU to excel at inference.</p><p>The new inference chip underscores the ongoing trend of AI adoption, which shifts the computational load from just building ever bigger models to actually using those models at scale. “Nvidia’s announcement validates the importance of SRAM-based architectures for large-scale inference, and no one has pushed SRAM density further than d-Matrix,” says d-Matrix CEO Sid Sheth. He’s betting that data center customers will want a variety of processors for inference. “The winning systems will combine different types of silicon and fit easily into existing data centers alongside GPUs.”</p><p>Inference-only chips may not be the only solution. Late last week, <a href="https://press.aboutamazon.com/aws/2026/3/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud" rel="noopener noreferrer" target="_blank">Amazon Web Services</a> said that it will deploy a new kind of inferencing system in its data centers. The system is a combination of AWS’s Tranium <a href="https://spectrum.ieee.org/amazon-ai" target="_self">AI accelerator </a>and <a href="https://spectrum.ieee.org/cerebras-chip-cs3" target="_self">Cerebras Systems’ third generation computer CS-3</a>, which is built around the <a href="https://spectrum.ieee.org/cerebrass-giant-chip-will-smash-deep-learnings-speed-barrier" target="_self">largest single chip</a> ever made. The two-part system is meant to take advantage of a technique called inference disaggregation. It separates inference into two parts—processing the prompt, called prefill, and generating the output, called decode. Prefill is inherently parallel, computationally intensive, and doesn’t need much memory bandwidth, while decode is a more serial process that needs a lot of memory bandwidth. Cerebras has maximized the memory bandwidth issue by building 44 GB of SRAM on its chip connected by a 21 PB/s network. </p><p><span>Nvidia, too, intends to take advantage of inference disaggregation in its new compute rack, called the Nvidia <a href="https://developer.nvidia.com/blog/inside-nvidia-groq-3-lpx-the-low-latency-inference-accelerator-for-the-nvidia-vera-rubin-platform/" target="_blank">Groq 3 LPX</a>. Each tray within the rack will house 8 Groq 3 LPUs. The LPX will split the inference task with a <a href="https://www.nvidia.com/en-us/data-center/vera-rubin-nvl72/" rel="noopener noreferrer" target="_blank">Vera Rubin NVL72</a>, Nvidia’s existing GPU and CPU rack.</span> The prefill and the more computationally intensive parts of the decode are done on Vera Rubin, while the final part is done on the Groq 3 LPU, leveraging the strengths of each chip. “We’re in volume production now,” Huang said.</p><p><br/></p><p><strong>Correction on 4/8/26: </strong>a previous version of this article incorrectly stated that the Nvidia Groq 3 LPX contains a Vera Rubin chip in each tray. In fact, each tray contains 8 Groq 3 LPUs and no Vera Rubins, but the whole rack is designed to work in concert with an NVL72 rack, which houses Vera Rubin chips. </p>]]></description><pubDate>Mon, 16 Mar 2026 21:04:33 +0000</pubDate><guid>https://spectrum.ieee.org/nvidia-groq-3</guid><category>Inferencing</category><category>Nvidia</category><category>Gpus</category><category>Processors</category><category>Ai</category><dc:creator>Dina Genkina</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-man-in-all-black-presents-in-front-of-a-large-screen-which-compares-a-large-rectangular-chip-labelled-rubin-gpu-with-a-square.jpg?id=65298681&amp;width=980"></media:content></item><item><title>Intel Demos Chip to Compute With Encrypted Data</title><link>https://spectrum.ieee.org/fhe-intel</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/overhead-view-of-intel-s-computing-chip-called-heracles.jpg?id=65174073&width=1200&height=800&coordinates=156%2C0%2C156%2C0"/><br/><br/><div class="ieee-summary"><h2>Summary</h2><ul><li><a href="#fhe">Fully homomorphic encryption (FHE)</a> allows computing on encrypted data without decryption, but it’s currently slow on standard CPUs and GPUs.</li><li>Intel’s Heracles chip accelerates FHE tasks up to <a href="#faster">5,000 times as fast as</a> top Intel server CPUs.</li><li>Heracles uses a <a href="#heracles">3-nanometer FinFET technology and high-bandwidth memory</a>, enabling efficient encrypted computing at scale.</li><li>Startups and Intel are <a href="#commercial">racing to commercialize FHE accelerators</a>, with potential applications in AI and secure data processing.</li></ul></div><p><span>Worried that your latest ask to a cloud-based AI reveals a bit too much about you? Want to know your genetic risk of disease without revealing it to the services that compute the answer?</span></p><p>There is a way to do computing on encrypted data without ever having it decrypted. It’s called <a href="https://spectrum.ieee.org/homomorphic-encryption" target="_blank">fully homomorphic encryption,</a> or FHE. But there’s a rather large catch. It can take thousands—even tens of thousands—of times as long to compute on today’s CPUs and GPUs than simply working with the decrypted data.</p><p>So universities, startups, and at least one processor giant have been working on specialized chips that could close that gap. Last month at the <a href="https://www.isscc.org/" target="_blank">IEEE International Solid-State Circuits Conference</a> (ISSCC) in San Francisco, <a href="https://www.intel.com/content/www/us/en/homepage.html" target="_blank">Intel</a> demonstrated its answer, Heracles, which sped up FHE computing tasks as much as 5,000-fold compared to a top-of the-line Intel server CPU.</p><p>Startups are racing to beat Intel and each other to commercialization. But <a href="https://www.linkedin.com/in/sanu-mathew-4073742/" target="_blank">Sanu Mathew,</a> who leads security circuits research at Intel, believes the CPU giant has a big lead, because its chip can do more computing than any other FHE accelerator yet built. “Heracles is the first hardware that works at scale,” he says.</p><p>The scale is measurable both physically and in compute performance. While other FHE research chips have been in the range of 10 square millimeters or less, Heracles is about 20 times that size and is built using Intel’s most advanced, 3-nanometer FinFET technology. And it’s flanked inside a liquid-cooled package by two 24-gigabyte <a href="https://spectrum.ieee.org/dram-shortage" target="_blank">high-bandwidth memory </a>chips—a configuration usually seen only in GPUs for training AI.</p><p class="ieee-inbody-related">RELATED: <a href="https://spectrum.ieee.org/how-to-compute-with-data-you-cant-see" target="_blank">How to Compute with Data You Can’t See</a></p><p>In terms of scaling compute performance, Heracles showed muscle in live demonstrations at ISSCC. At its heart the demo was a simple private query to a secure server. It simulated a request by a voter to make sure that her ballot had been registered correctly. The state, in this case, has an encrypted database of voters and their votes. To maintain her privacy, the voter would not want to have her ballot information decrypted at any point; so using FHE, she encrypts her ID and vote and sends it to the government database. There, without decrypting it, the system determines if it is a match and returns an encrypted answer, which she then decrypts on her side.</p><p>On an Intel Xeon server CPU, the process took 15 milliseconds. Heracles did it in 14 microseconds. While that difference isn’t something a single human would notice, verifying 100 million voter ballots adds up to more than 17 days of CPU work versus a mere 23 minutes on Heracles.</p><p>Looking back on the five-year journey to bring the Heracles chip to life, <a href="https://www.linkedin.com/in/ro-cammarota-a226b817/" target="_blank">Ro Cammarota</a>, who led the project at Intel until last December and is now at University of California, Irvine, says “We have proven and delivered everything that we promised.”</p><h2>FHE Data Expansion</h2><p class="rm-anchors" id="fhe">FHE is fundamentally a mathematical transformation, sort of like the Fourier transform. It encrypts data using a quantum-computer-proof algorithm, but, crucially, uses corollaries to the mathematical operations usually used on unencrypted data. These corollaries achieve the same ends on the encrypted data.<strong></strong></p><p>One of the main things holding such secure computing back is the explosion in the size of the data once it’s encrypted for FHE, <a href="https://www.linkedin.com/in/anupamgolder/" target="_blank">Anupam Golder</a>, a research scientist at Intel’s circuits research lab, told engineers at ISSCC. “Usually, the size of cipher text is the same as the size of plain text, but for FHE it’s orders of magnitude larger,” he said.</p><p>While the sheer volume is a big problem, the kinds of computing you need to do with that data is also an issue. FHE is all about very large numbers that must be computed with precision. While a CPU can do that, it’s very slow going—integer addition and multiplication take about 10,000 more clock cycles in FHE. Worse still, CPUs aren’t built to do such computing in parallel. Although GPUs excel at parallel operations, precision is not their strong suit. (In fact, from generation to generation, GPU designers have devoted more and more of the chip’s resources to <a href="https://spectrum.ieee.org/nvidia-gpu" target="_blank">computing less-and-less-precise numbers</a>.)</p><p>FHE also requires some oddball operations with names like “twiddling” and “automorphism,” and it relies on a compute-intensive noise-cancelling process called bootstrapping. None of these things are efficient on a general-purpose processor. So, while clever algorithms and libraries of software cheats have been developed over the years, the need for a hardware accelerator remains if FHE is going to tackle large-scale problems, says Cammarota.</p><h2>The Labors of Heracles</h2><p class="rm-anchors" id="heracles">Heracles was initiated under a <span>Defense Advanced Research Projects Agency</span> (DARPA) program five years ago to accelerate FHE using purpose-built hardware. It was developed as “a whole system-level effort that went all the way from theory and algorithms down to the circuit design,” says Cammarota.</p><p>Among the first problems was how to compute with numbers that were larger than even the 64-bit words that are today a CPU’s most precise. There are ways to break up these gigantic numbers into chunks of bits that can be calculated independently of each other, providing a degree of parallelism. Early on, the Intel team made a big bet that they would be able to make this work in smaller, 32-bit chunks, yet still maintain the needed precision. This decision gave the Heracles architecture some speed and parallelism, because the 32-bit arithmetic circuits are considerably smaller than 64-bit ones, explains Cammarota.</p><p>At Heracles’s heart are 64 compute cores—called tile-pairs—arranged in an eight-by-eight grid. These are what are called single instruction multiple data (SIMD) compute engines designed to do the polynomial math, twiddling, and other things that make up computing in FHE and to do them in parallel. An on-chip 2D mesh network connects the tiles to each other with wide, 512-byte, buses.</p><p class="ieee-inbody-related">RELATED: <a href="https://spectrum.ieee.org/homomorphic-encryption-llm" target="_blank">Tech Keeps Chatbots From Leaking Your Data</a></p><p>Important to making encrypted computing efficient is feeding those huge numbers to the compute cores quickly. The sheer amount of data involved meant linking 48-GB-worth of expensive high-bandwidth memory to the processor with 819-GB-per-second connections. Once on the chip, data musters in 64 megabytes of cache memory—somewhat more than an Nvidia <a href="https://spectrum.ieee.org/nvidias-next-gpu-shows-that-transformers-are-transforming-ai" target="_blank">Hopper-generation GPU</a>. From there it can flow through the array at 9.6 terabytes per second by hopping from tile-pair to tile-pair.</p><p>To ensure that computing and moving data don’t get in each other’s way, Heracles runs three synchronized streams of instructions simultaneously, one for moving data onto and off of the processor, one for moving data within it, and a third for doing the math, Golder explained.</p><p class="rm-anchors" id="faster">It all adds up to some massive speedups, according to Intel. Heracles—operating at 1.2 gigahertz—takes just 39 microseconds to do FHE’s critical math transformation, a 2,355-fold improvement over an Intel Xeon CPU running at 3.5 GHz. Across seven key operations, Heracles was 1,074 to 5,547 times as fast.</p><p>The differing ranges have to do with how much data movement is involved in the operations, explains Mathew. “It’s all about balancing the movement of data with the crunching of numbers,” he says.</p><h2>FHE Competition</h2><p class="rm-anchors" id="commercial">“It’s very good work,” <a href="https://www.linkedin.com/in/kurt-rohloff/" target="_blank">Kurt Rohloff</a>, chief technology officer at FHE software firm <a href="https://dualitytech.com/platform/technology-fully-homomorphic-encryption/" target="_blank">Duality Technology</a>, says of the Heracles results. Duality was part of a team that developed a competing accelerator design under the same DARPA program that brought forth Intel’s Heracles. “When Intel starts talking about scale, that usually carries quite a bit of weight.”</p><p>Duality’s focus is less on new hardware than on software products that do the kind of encrypted queries that Intel demonstrated at ISSCC. At the scale in use today “there’s less of a need for [specialized] hardware,” says Rohloff. “Where you start to need hardware is emerging applications around deeper machine-learning-oriented operations like neural net, LLMs, or semantic search.”</p><p>Last year, Duality demonstrated an <a href="https://spectrum.ieee.org/homomorphic-encryption-llm" target="_self">FHE-encrypted language model called BERT</a>. Like more famous LLMs such as ChatGPT, BERT is a transformer model. However it’s only one-tenth the size of even the most compact LLMs.</p><p><a href="https://www.linkedin.com/in/barrus/" target="_blank">John Barrus</a>, vice president of product at Dayton, Ohio–based <a href="https://niobiummicrosystems.com/" target="_blank">Niobium Microsystems</a>, an FHE chip startup <a href="https://www.galois.com/" target="_blank">spun out</a> of another DARPA competitor, agrees that encrypted AI is a key target of FHE chips. “There are a lot of smaller models that, even with FHE’s data expansion, will run just fine on accelerated hardware,” he says.</p><p>With no stated commercial plans from Intel, Niobium expects its chip to be “the world’s first commercially viable FHE accelerator, designed to enable encrypted computations at speeds practical for real-world cloud and AI infrastructure.” Although it hasn’t announced when a commercial chip will be available, last month the startup revealed that it had inked a deal worth 10 billion South Korean won (US $6.9 million) with Seoul-based chip design firm <a href="https://semifive.com/" target="_blank">Semifive</a> to develop the FHE accelerator for fabrication using Samsung Foundry’s 8-nanometer-process technology.</p><p>Other startups including <a href="https://cornami.com/" target="_blank">Cornami</a>,  <a href="https://www.fabriccryptography.com/" target="_blank">Fabric Cryptography</a>, and <a href="https://optalysys.com/" target="_blank">Optalysys</a> have been working on chips to accelerate FHE. Optalysys CEO <a href="https://optalysys.com/people/" target="_blank">Nick New</a> says Heracles hits about the level of speedup you could hope for using an all-digital system. “We’re looking at pushing way past that digital limit,” he says. His company’s approach is to use the physics of a photonic chip to do FHE’s compute-intensive transform steps. That photonics chip is on its seventh generation, he says, and among the next steps is to 3D integrate it with custom silicon to do the nontransform steps and coordinate the whole process. A full 3D-stacked commercial chip could be ready in two or three years, says New.</p><p>While competitors develop their chips, so will Intel, says Mathew. It will be improving on how much the chip can accelerate computations by fine-tuning the software. It will also be trying out more massive FHE problems, and exploring hardware improvements for a potential next generation. “This is like the first microprocessor…the start of a whole journey,” says Mathew.</p>]]></description><pubDate>Tue, 10 Mar 2026 13:00:04 +0000</pubDate><guid>https://spectrum.ieee.org/fhe-intel</guid><category>Privacy</category><category>Intel</category><category>Encryption</category><category>Homomorphic-encryption</category><category>Hardware-acceleration</category><category>Isscc</category><dc:creator>Samuel K. Moore</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/overhead-view-of-intel-s-computing-chip-called-heracles.jpg?id=65174073&amp;width=980"></media:content></item><item><title>Finite-Element Approaches to Transformer Harmonic and Transient Analysis</title><link>https://content.knowledgehub.wiley.com/solving-harmonic-and-transient-challenges-in-transformers-using-integrateds-faraday/</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/logo-of-integrated-engineering-software-with-pixelated-geometric-design-and-text.png?id=65106417&width=980"/><br/><br/><p>Explore structured finite-element methodologies for analyzing transformer behavior under harmonic and transient conditions — covering modelling, solver configuration, and result validation techniques.</p><p><strong>What Attendees will Learn</strong><span></span></p><ol><li>How FEM enables pre-fabrication performance evaluation — Assess magnetic field distribution, current behavior, and turns-ratio accuracy through simulation rather than physical testing.</li><li><span>How harmonic analysis uncovers saturation and imbalance — Identify high-flux regions and current asymmetries that analytical methods may not capture.</span></li><li><span>How transient simulations characterize dynamic response — Examine time-domain current waveforms, inrush behavior, and multi-cycle stabilization.</span></li><li><span>How modelling choices affect simulation fidelity — Understand the impact of coil definitions, winding configurations, solver type, and material models on accuracy.</span></li></ol><p><span><a href="https://content.knowledgehub.wiley.com/solving-harmonic-and-transient-challenges-in-transformers-using-integrateds-faraday/" target="_blank">Download this free whitepaper now!</a><br/></span></p>]]></description><pubDate>Tue, 10 Mar 2026 10:00:03 +0000</pubDate><guid>https://content.knowledgehub.wiley.com/solving-harmonic-and-transient-challenges-in-transformers-using-integrateds-faraday/</guid><category>Type-whitepaper</category><category>Transformers</category><category>Finite-element-analysis</category><category>Harmonic</category><dc:creator>Integrated Engineering Software</dc:creator><media:content medium="image" type="image/png" url="https://assets.rbl.ms/65106417/origin.png"></media:content></item><item><title>Entomologists Use a Particle Accelerator to Image Ants at Scale</title><link>https://spectrum.ieee.org/3d-scanning-particle-accelerator-antscan</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/four-grey-3d-models-of-ants-shown-up-close-in-high-detail-two-larger-ants-tower-above-two-smaller-ones-in-the-front-the-larges.jpg?id=65150255&width=1200&height=800&coordinates=62%2C0%2C63%2C0"/><br/><br/><p>Move over, Pixar. The ants that animators once morphed into googly-eyed caricatures in films such as <em>A Bug’s Life</em> and <em>Antz</em> just received a meticulously precise anatomical reboot.</p><p><a href="https://doi.org/10.1038/s41592-026-03005-0" rel="noopener noreferrer" target="_blank">Writing today in <em>Nature Methods</em></a>, an international team of entomologists, accelerator physicists, computer scientists, and biological-imaging specialists describe a new 3D atlas of ant morphology.</p><p>Dubbed Antscan, the platform features micrometer-resolution reconstructions that lay bare not only the <a href="https://spectrum.ieee.org/festo-bionic-ants-and-butterflies" target="_self">insects’ armored exoskeletons</a> but also their muscles, nerves, digestive tracts, and needlelike stingers poised at the ready.</p><p>Those high-resolution images—spanning 792 species across 212 genera and covering the bulk of described ant diversity—are now available free of charge through an <a href="http://www.antscan.info" rel="noopener noreferrer" target="_blank">interactive online portal</a>, where anyone can rotate, zoom, and virtually “dissect” the insects from a laptop.</p><p>“Antscan is exciting!” says <a href="https://experts.mcmaster.ca/people/curric7" rel="noopener noreferrer" target="_blank">Cameron Currie</a>, an evolutionary biologist at McMaster University in Hamilton, Ont., Canada, who was not involved in the research. “It provides an outstanding resource for comparative work across ants.”</p><h2>Digital Access to Natural History Collections</h2><p>It also provides broader access to natural history collections.</p><p>No longer must these vast archives of preserved life be confined to drawers and jars in museums scattered around the world, available only to specialists able to visit in person. All these specimens can now be explored digitally by anyone with an internet connection, adding fresh scientific value to museum holdings.</p><p>“The more people that access and work with the stuff in our museums, whether it’s physically or digitally, the greater value they add,” says <a href="https://www.floridamuseum.ufl.edu/blackburn-lab/personnel/principal-investigator/" rel="noopener noreferrer" target="_blank">David Blackburn</a>, the curator of herpetology at the Florida Museum of Natural History who, like Currie, was not involved in the research.</p><p>Some of those people may be professional myrmecologists (scientists who specialize in the study of ants) and fourmiculture (ant-farming) enthusiasts. But others may be schoolteachers, video-game designers, tattoo artists, or curious members of the public.</p><p>“It is an extremely rich dataset that can be used for a number of different applications in science, but  also for the arts and outreach and education.” says <a href="https://www.oist.jp/image/julian-katzke" rel="noopener noreferrer" target="_blank"><span>Julian Katzke</span></a>, an entomologist at the National Museum of Natural History in Washington, D.C.</p><p>Card-carrying members of <em>IEEE</em> should find plenty to explore in Antscan as well, says <a href="https://entomology.umd.edu/people/evan-economo" target="_blank">Evan Economo</a>, a biodiversity scientist at the University of Maryland in College Park who, along with Katzke, co-led the project. <span>With the dataset now publicly available and standardized at scale, “I would really like to see these big libraries of organismal form one day be useful for people in robotics and engineering, so they can mine these data for new kinds of biomechanical designs,” he says.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Various 3D renderings of an ant soldier. First, the outward appearance. Followed by cross sectional slices of its body. One shows the internal structures of the ant, with space predominantly occupied by muscles. Another shows the same view, but with muscles removed, which highlights the digestive tract and nervous system. Lastly, zoomed-in renderings inside the ant's brain, gut and sting apparatus are shown with labels." class="rm-shortcode" data-rm-shortcode-id="672fbae791e49ff86839c1593eccc48d" data-rm-shortcode-name="rebelmouse-image" id="33b51" loading="lazy" src="https://spectrum.ieee.org/media-library/various-3d-renderings-of-an-ant-soldier-first-the-outward-appearance-followed-by-cross-sectional-slices-of-its-body-one-show.jpg?id=65150295&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">These renderings reveal different structures within the body of an army ant (<i>Eciton hamatum</i>) subsoldier, based on Antscan data.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit..."><a href="https://doi.org/10.1038/s41592-026-03005-0" target="_blank">Katzke et al.</a></small></p><h2>Advancements in Ant Imaging Technology</h2><p>Researchers have been digitizing natural history collections for years: photographing drawers of pinned specimens, building surface-level models from overlapping image stacks, and using computed tomography (CT) to scan select species one at a time. But those efforts are typically slow, piecemeal, and often limited to external features.</p><p>To capture entire organisms, inside and out, Economo and his team—then based at the Okinawa Institute of Science and Technology in Japan and including former lab members Katzke and <a href="https://www.museumfuernaturkunde.berlin/en/research/research/dynamics-nature/center-integrative-biodiversity-discovery" target="_blank">Francisco Hita Garcia</a> (now at the Museum für Naturkunde in Berlin)—built an automated imaging pipeline that effectively turned a particle accelerator into an anatomical assembly line.</p><p>They scoured museum collections around the world for ant specimens—workers, queens, and males alike—and sent some 2,200 preserved samples through a pair of micro-CT beamlines at the Karlsruhe Institute of Technology’s synchrotron <a href="https://www.ibpt.kit.edu/KIT_Light_Source.php" target="_blank">light source facility</a> in Germany.<strong></strong></p><p>There, biological imaging specialist <a href="https://www.ips.kit.edu/2890_5177.php" target="_blank">Thomas van de Kamp</a> oversaw the operation, as intense X-ray beams swept through each specimen and high-speed detectors recorded thousands of projection images from multiple angles. Robotic handlers moved vials of alcohol-preserved ants into position, one after another, all in a matter of days.</p><p>Software then reconstructed 200-plus terabytes of data generated into 3D volumes, with neural networks helping to automate the identification and analysis of anatomical structures.</p><p>Similar large-scale digitization efforts—such as the <a href="https://www.floridamuseum.ufl.edu/overt/" target="_blank">openVertebrate Project</a>, led by the Florida Museum of Natural History’s Blackburn, which involved <a href="https://academic.oup.com/bioscience/article/74/3/169/7615104" target="_blank">scanning thousands</a> of birds, fish, mammals, reptiles, and amphibians—have begun transforming how biologists study anatomy. But applying conventional micro-CT at comparable scale to insects, which are smaller and harder to scan at useful resolutions, required a leap in speed and throughput.</p><p>That’s where the synchrotron came in. By harnessing a particle accelerator to generate extraordinarily bright, coherent X-rays, the team was able to capture high-resolution internal anatomy in seconds, without the lengthy staining or other preprocessing steps often required for soft-tissue contrast in standard lab scanners.</p><p>“It is an impressive piece of work,” says <a href="https://www.nms.ac.uk/profile/dr-vladimir-blagoderov" target="_blank">Vladimir Blagoderov</a>, principal curator of invertebrates at the National Museums Scotland in Edinburgh, who was not involved in the research. “This project adds an industrial dimension to CT scanning by combining robotics, standardized sampling, automated image-processing pipelines, and machine learning.”</p><p>The sheer taxonomic breadth of the Antscan dataset now makes it possible to spot patterns across the entire ant family tree, as Economo and his colleagues have already demonstrated.</p><p>In a separate paper published last December, for example, the researchers drew on the newly generated scans to measure how much ants invest in their outer protective casing. Reporting in <em>Science Advances,</em> they showed that species with lighter, less costly cuticles <a href="https://www.science.org/doi/10.1126/sciadv.adx8068" target="_blank"><span><span>tend to form larger colonies and diversify more rapidly</span></span></a> over evolutionary time.</p><p>In their latest study, the Antscan team  turned to a different evolutionary question: The distribution of a biomineral “armor” layer <a href="https://www.nature.com/articles/s41467-020-19566-3" target="_blank">first described</a> by Currie and his colleagues in 2020 in a Central American leaf-cutter ant. A quick sweep through the Antscan database revealed that this coating—which absorbs X-rays and is visible as a bright sheath over the cuticle—is not an oddity confined to one species.</p><p>Instead, it is common among fungus-farming ants, the evolutionary lineage from which leaf-cutting ants arose roughly 20 million years ago, but largely absent in most other branches of the ant tree. (Currie’s team independently confirmed the pattern using X-ray diffraction, a technique that can precisely reveal a material’s mineral composition, as the group <a href="https://www.biorxiv.org/content/10.64898/2026.02.07.704540v1" target="_blank">reported last month in a preprint</a> posted to <em>bioRxiv</em>.)</p><p>Those are only early demonstrations of what the database can do, though. And with AI tools increasingly capable of parsing enormous, information-rich data troves, the real analytical power of Antscan may still lie ahead, says <a href="https://agsci.colostate.edu/agbio/gillette-museum/museum-staff/" target="_blank">Marek Borowiec</a>, director of the C.P. Gillette Museum of Arthropod Diversity at Colorado State University, who has chronicled <a href="https://besjournals.onlinelibrary.wiley.com/doi/full/10.1111/2041-210X.13901" target="_blank">the rise of <span>deep learning tools</span></a> in ecology and evolution.</p><p>“The full advantage of this dataset will be realized when these methods are deployed,” he says.</p><h2>Transforming Morphology with Antscan</h2><p>The ambitions behind Antscan extend well beyond ant biology. Economo and his colleagues see it as a blueprint for digitizing, standardizing, and scaling anatomy itself.<br/></p><p>Just as <a href="https://spectrum.ieee.org/whole-genome-sequencing" target="_self">large-scale sequencing projects</a> and genomic databases transformed the study of DNA over the past two decades, they hope Antscan will catalyze a comparable shift for morphology. <span>“This is kind of like having a genome for shape,” Economo says.</span></p><p>Museum collections house millions of alcohol-preserved insects and other small invertebrates—beetles, flies, wasps, spiders, crustaceans—many of them representing rare or extinct populations. Following the Antscan playbook, each could be converted into a high-resolution library of “<a data-linked-post="2655774779" href="https://spectrum.ieee.org/climate-models" target="_blank">digital twins.</a>“</p><p>In each case, synchrotron micro-CT would offer a rapid way to peer inside fragile specimens without cutting them open, capturing both hard exoskeleton and soft tissue in exquisite detail across vast swaths of biological diversity.</p><p class="shortcode-media shortcode-media-youtube"> <span class="rm-shortcode" data-rm-shortcode-id="a20837327321eee6ad3fab098e4da2e3" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/neYh_KITjGE?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span> <small class="image-media media-photo-credit" placeholder="Add Photo Credit..."><a href="https://www.youtube.com/watch?v=neYh_KITjGE" target="_blank">Antscan/YouTube</a></small></p><p><span>Beam time at major synchrotron facilities is scarce and fiercely competitive, a practical bottleneck for any effort to digitize biodiversity at scale, notes National Museums Scotland’s Blagoderov. What’s more, “even once the scans exist, the downstream burden is nontrivial: M</span><span>oving, storing, and processing hundreds of terabytes of data can become a bottleneck in its own right,” he says.</span></p><p>But if access can be secured and the computational infrastructure scaled to match, such efforts could transform natural history museums from static repositories into dynamic digital biomes.</p><p>That transformation may prove especially important at a time of accelerating species loss on Earth. By capturing organisms in extraordinary detail, resources like Antscan create a permanent, high-resolution record of life’s architecture—an anatomical time capsule that can be queried and revisited long after fragile specimens degrade or wild populations vanish.</p><p>And should Pixar ever greenlight <em>A Bug’s Life 2 </em>(suggested title: <em>Even Buggier</em>),<em> </em>the studio’s character designers may not need to take much artistic license at all. Thanks to a particle accelerator and a small cadre of dedicated scientists, the reference models are already in hand—rendered not in animation software but in micrometer-perfect anatomical form.</p>]]></description><pubDate>Thu, 05 Mar 2026 10:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/3d-scanning-particle-accelerator-antscan</guid><category>Machine-learning</category><category>Insects</category><category>Particle-accelerator</category><category>Computed-tomography</category><dc:creator>Elie Dolgin</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/four-grey-3d-models-of-ants-shown-up-close-in-high-detail-two-larger-ants-tower-above-two-smaller-ones-in-the-front-the-larges.jpg?id=65150255&amp;width=980"></media:content></item><item><title>Watershed Moment for AI–Human Collaboration in Math</title><link>https://spectrum.ieee.org/ai-proof-verification</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/four-by-four-grid-of-circles-with-varying-color-gradient-patterns.jpg?id=65103143&width=2000&height=1500&coordinates=0%2C0%2C0%2C0"/><br/><br/><p><span>When Ukrainian mathematician </span><a href="https://people.epfl.ch/maryna.viazovska?lang=en" target="_blank">Maryna Viazovska</a><span> received a </span><a href="https://www.mathunion.org/imu-awards/fields-medal/fields-medals-2022" target="_blank">Fields Medal</a><span>—widely regarded as the Nobel Prize for mathematics—in July 2022,</span><span> it was big news. Not only was she the second woman to accept the honor in the award’s 86-year history, but she collected the medal just months after her country had been invaded by Russia. Nearly four years later, Viazovska is making waves again. <a href="https://www.math.inc/sphere-packing" target="_blank">Today</a>, in </span><span>a collaboration between humans and AI, Viazovska’s proofs have been formally verified, signaling rapid progress in AI’s abilities to <a href="https://spectrum.ieee.org/ai-math-benchmarks" target="_blank">assist</a> with mathemat</span><span>ical research. </span></p><p><span>“These new results seem very, very impressive, and definitely signal some rapid progress in this direction,” says AI-reasoning expert and Princeton University postdoc <a href="https://ai.princeton.edu/news/2025/ai-lab-welcomes-associate-research-scholars" target="_blank">Liam Fowl</a>, who was not involved in the work.</span></p><p>In her Fields Medal–winning research, Viazovska had tackled two versions of the sphere-packing problem, which asks: How densely can identical circles, spheres, et cetera, be packed in <em>n</em>-dimensional space? In two dimensions, the honeycomb is the best solution. In three dimensions, spheres stacked in a pyramid are optimal. But after that, it becomes exceedingly difficult to find the best solution, and to prove that it is in fact the best. </p><p>In 2016, Viazovska solved the problem in two cases. By using powerful mathematical functions known as (quasi-)modular forms, she proved that a symmetric arrangement known as E<sub>8</sub> is the <a href="https://annals.math.princeton.edu/articles/keyword/sphere-packing" target="_blank">best 8-dimensional packing</a>, and soon after proved with collaborators that another sphere packing called the <a href="https://annals.math.princeton.edu/2017/185-3/p08" target="_blank">Leech lattice is best in 24 dimensions</a>. Though seemingly abstract, this result has potential to help solve everyday problems related to dense sphere packing, including <a data-linked-post="2650280110" href="https://spectrum.ieee.org/novel-error-correction-code-opens-a-new-approach-to-universal-quantum-computing" target="_blank">error-correcting codes</a> used by smartphones and space probes.</p><p>The proofs were verified by the mathematical community and deemed correct, leading to the Fields Medal recognition. But formal verification—the ability of a proof to be verified by a computer—is another beast altogether. Since 2022, much <a href="https://cacm.acm.org/research/formal-reasoning-meets-llms-toward-ai-for-mathematics-and-verification/" target="_blank">progress</a> has been made in AI-assisted formal proof verification. </p><h2>Serendipity leads to formalization project</h2><p>A few years later, a chance meeting in Lausanne, Switzerland, between third-year undergraduate <a href="https://thefundamentaltheor3m.github.io/" target="_blank">Sidharth Hariharan</a> and Viazovska would reignite her interest in sphere-packing proofs. Though still very early in his career, Hariharan was already becoming adept at formalizing proofs.</p><p>“Formal verification of a proof is like a rubber stamp,” Fowl says. “It’s a kind of bona fide certification that you know your statements of reasoning are correct.”</p><p>Hariharan told Viazovska how he had been using the process of formalizing proofs to learn and really understand mathematical concepts. In response, Viazovska expressed an interest in formalizing her proofs, largely out of curiosity. From this, in March 2024 the <a href="https://thefundamentaltheor3m.github.io/Sphere-Packing-Lean/" target="_blank">Formalising Sphere Packing in Lean</a> project was born. <span>Lean is a popular programming language and “proof assistant” that allows mathematicians to write proofs that are then verified for absolute correctness by a computer.</span></p><p>A collaboration formed to write a human-readable “blueprint” that could be used to map the 8-dimensional proof’s various constituents and figure out which of them had and had not been formalized and/or proven, and then prove and formalize those missing elements in Lean. </p><p><span>“We had been building the project’s repository for about 15 months when we enabled public access in June 2025,” recalls Hariharan, now a first-year Ph.D. student at Carnegie Mellon University. “Then, in late October we heard from Math, Inc. for the first time.”</span></p><h2>The AI speedup</h2><p><a href="https://www.math.inc/" target="_blank">Math, Inc.</a> is a startup developing Gauss, an AI specifically designed to automatically formalize proofs. “It’s a particular kind of language model called a reasoning agent that’s meant to interleave both traditional natural-language reasoning and fully formalized reasoning,” explains <a href="https://jesse-michael-han.github.io/" target="_blank">Jesse Han</a>, Math, Inc. CEO and cofounder. “So it’s able to conduct literature searches, call up tools, and use a computer to write down Lean code, take notes, spin up verification tooling, run the Lean compiler, et cetera.”</p><p>Math, Inc. first hit the headlines when it announced that Gauss had completed a <a href="https://mathstodon.xyz/@tao/111847680248482955" target="_blank">Lean formalization of the strong <span>prime number theorem</span> (PNT)</a> in three weeks last summer, a task that Fields Medalist <a href="https://terrytao.wordpress.com/" target="_blank">Terence Tao</a> and <a href="https://sites.math.rutgers.edu/~alexk/" target="_blank">Alex Kontorovich</a> had been working on. Similarly, Math, Inc. contacted Hariharan and colleagues to say that Gauss had proven several facts related to their sphere-packing project.</p><p>“They told us that they had finished 30 “sorrys,” which meant that they proved 30 intermediate facts that we wanted proved,” explains Hariharan. A proportion of these sorrys were shared with the project team and merged with their own work. “One of them helped us identify a typo in our project, which we then fixed,” adds Hariharan. “So it was a pretty fruitful collaboration.”</p><h2>From 8 to 24 dimensions</h2><p>But then, radio silence followed. Math, Inc. appeared to lose interest. However, while Hariharan and colleagues continued their labor of love, Math, Inc. was building a new and improved version of Gauss. “We made a research breakthrough sometime mid-January that produced a much stronger version of Gauss,” says Han. “This new version reproduced our three-week PNT result in two to three days.”</p><p>Days later, the new Gauss was steered back to the sphere-packing formalization. Working from the invaluable preexisting blueprint and work that Hariharan and collaborators had shared, Gauss not only autoformalized the 8-dimensional case, but also found and fixed a typo in the published paper, all in the space of five days.</p><p>“When they reached out to us in late January saying that they finished it, to put it very mildly, we were very surprised,” says Hariharan. “But at the end of the day, this is technology that we’re very excited about, because it has the capability to do great things and to assist mathematicians in remarkable ways.”</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="A laptop with sphere packing code in the foreground, with an autumn sunset at Carnegie Mellon in the background. " class="rm-shortcode" data-rm-shortcode-id="1dd0742602809b330ce11552ae9d6d3f" data-rm-shortcode-name="rebelmouse-image" id="898fd" loading="lazy" src="https://spectrum.ieee.org/media-library/a-laptop-with-sphere-packing-code-in-the-foreground-with-an-autumn-sunset-at-carnegie-mellon-in-the-background.jpg?id=65106120&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Hariharan was working on sphere-packing proof verification as the sun was setting behind Carnegie Mellon’s Hamerschlag Hall.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Sidharth Hariharan</small></p><p>The 8-dimensional sphere-packing proof formalization alone, <a href="https://leanprover.zulipchat.com/#narrow/channel/113486-announce/topic/Sphere.20Packing.20Milestone/with/575354368" target="_blank">announced on February 23</a>, represents a watershed moment for autoformalization and AI–human collaboration. But <a target="_blank"></a><a href="https://math.inc/sphere-packing" target="_blank">today, Math, Inc. revealed</a><span> </span>an even more impressive accomplishment: Gauss has autoformalized Viazovska’s 24-dimensional sphere-packing proof—all 200,000+ lines of code of it—in just two weeks. </p><p>There are commonalities between the 8- and 24-dimensional cases in terms of the foundational theory and overall architecture of the proof, meaning some of the code from the 8-dimensional case could be refactored and reused. However, Gauss had no preexisting blueprint to work from this time. “And it was actually significantly more involved than the 8-dimensional case, because there was a lot of missing background material that had to be brought on line surrounding many of the properties of the Leech lattice, in particular its uniqueness,” explains Han.</p><p>Though the 24-dimensional case was an automated effort, both Han and Hariharan acknowledge the many contributions from humans that laid the foundations for this achievement, regarding it as a collaborative endeavor overall between humans and AI.</p><p>But for Han, it represents even more: the beginning of a revolutionary transformation in mathematics, where extremely large-scale formalizations are commonplace. “A programmer used to be someone who punched holes into cards, but then the act of programming became separated from whatever material substrate was used for recording programs,” he concludes. “I think the end result of technology like this will be to free mathematicians to do what they do best, which is to dream of new mathematical worlds.”</p>]]></description><pubDate>Mon, 02 Mar 2026 18:00:03 +0000</pubDate><guid>https://spectrum.ieee.org/ai-proof-verification</guid><category>Mathematics</category><category>Ai-reasoning</category><category>Large-language-models</category><category>Ai</category><dc:creator>Benjamin Skuse</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/four-by-four-grid-of-circles-with-varying-color-gradient-patterns.jpg?id=65103143&amp;width=980"></media:content></item><item><title>How Quantum Data Can Teach AI to Do Better Chemistry</title><link>https://spectrum.ieee.org/quantum-chemistry</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/illustration-of-a-human-head-in-profile-with-a-spiral-upon-which-human-figures-are-walking-overlaid-on-an-image-of-an-atom.png?id=63744636&width=2000&height=1500&coordinates=0%2C183%2C0%2C184"/><br/><br/><p><strong>Sometimes a visually compelling</strong> metaphor is all you need to get an otherwise complicated idea across. In the summer of 2001, a Tulane physics professor named <a href="https://sse.tulane.edu/john-p-perdew-phd" rel="noopener noreferrer" target="_blank">John P. Perdew</a> came up with a banger. He wanted to convey the hierarchy of computational complexity inherent in the behavior of electrons in materials. He called it “<a href="https://pubs.aip.org/aip/acp/article-abstract/577/1/1/573973/Jacob-s-ladder-of-density-functional?redirectedFrom=fulltext" rel="noopener noreferrer" target="_blank">Jacob’s Ladder</a>.” He was appropriating an idea from the Book of Genesis, in which Jacob dreamed of a ladder “set up on the earth, and the top of it reached to heaven. And behold the angels of God ascending and descending on it.”</p><p>Jacob’s Ladder represented a gradient and so too did Perdew’s ladder, not of spirit but of computation. At the lowest rung, the math was the simplest and least computationally draining, with materials represented as a smoothed-over, cartoon version of the atomic realm. As you climbed the ladder, using increasingly more intensive mathematics and compute power, descriptions of atomic reality became more precise. And at the very top, nature was perfectly described via impossibly intensive computation—something like what God might see.</p><p>With this metaphor in mind, we propose to extend Jacob’s Ladder beyond Perdew’s version, to encompass <em><em>all</em></em> computational approaches to simulating the behavior of electrons. And instead of climbing rung by rung toward an unreachable summit, we have an idea to <em><em>bend</em></em> the ladder so that even the very top lies within our grasp. Specifically, we at Microsoft envision a hybrid approach. It starts with using quantum computers to generate exquisitely accurate data about the behavior of electrons—data that would be prohibitively expensive to compute classically. This quantum-generated data will then train AI models running on classical machines, which can predict the properties of materials with remarkable speed. By combining quantum accuracy with AI-driven speed, we can ascend Jacob’s Ladder faster, designing new materials with novel properties and at a fraction of the cost.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Graph comparing computational cost and simulation accuracy: Classical, DFT, Coupled, Quantum+AI." class="rm-shortcode" data-rm-shortcode-id="d3175e47f1efce66722968991732929d" data-rm-shortcode-name="rebelmouse-image" id="13461" loading="lazy" src="https://spectrum.ieee.org/media-library/graph-comparing-computational-cost-and-simulation-accuracy-classical-dft-coupled-quantum-ai.png?id=65172435&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">At the base of Jacob’s Ladder are classical models that treat atoms as simple balls connected by springs—fast enough to handle millions of atoms over long times but with the lowest precision. Moving up along the black line, semiempirical methods add some quantum mechanical calculations. Next are approximations based on Hartree-Fock (HF) and density functional theory (DFT), which include full quantum behavior of individual electrons but model their interactions in an averaged way. The greater accuracy requires significant computing power, which limits them to simulating molecules with no more than a few hundred atoms. At the top are coupled-cluster and full configuration interaction (FCI) methods—exquisitely accurate but, at the moment, restricted to tiny molecules or subsets of electrons due to the large computational costs involved. Quantum computing can bend the accuracy-versus-cost curve at the top [orange line], making highly accurate calculations feasible for large systems. AI, trained on this quantum-accurate data, can flatten this curve [purple line], enabling rapid predictions for similar systems at a fraction of the cost of classical computing.</small></p><p>In our approach, the base of Jacob’s Ladder still starts with classical models that treat atoms as simple balls connected by springs—models that are fast enough to handle millions of atoms over long times, but with the lowest precision. As we ascend the ladder, some quantum mechanical calculations are added to semiempirical methods. Eventually, we’ll get to the full quantum behavior of individual electrons but with their interactions modeled in an averaged way; this greater accuracy requires significant compute power, which means you can only simulate molecules of no more than a few hundred atoms. At the top will be the most computationally intensive methods—prohibitively expensive on classical computers but tractable on quantum computers.</p><p>In the coming years, quantum computing and AI will become critical tools in the pursuit of new materials science and chemistry. When combined, their forces will multiply. We believe that by using quantum computers to train AI on quantum data, the result will be hyperaccurate AI models that can reach ever higher rungs of computational complexity without the prohibitive computational costs.</p><p>This powerful combination of quantum computing and AI could unlock unprecedented advances in chemical discovery, materials design, and our understanding of complex reaction mechanisms. Chemical and materials innovations already play a vital—if often invisible—role in our daily lives. These discoveries shape the modern world: new drugs to help treat disease more effectively, improving health and extending life expectancy; everyday products like toothpaste, sunscreen, and cleaning supplies that are safe and effective; cleaner fuels and longer-lasting batteries; improved fertilizers and pesticides to boost global food production; and biodegradable plastics and recyclable materials to shrink our environmental footprint. In short, chemical discovery is a behind-the-scenes force that greatly enhances our everyday lives.</p><h3></h3><br/><div class="rblad-ieee_in_content"></div><p>The potential is vast. Anywhere AI is already in use, this new quantum-enhanced AI could drastically improve results. These models could, for instance, scan for previously unknown catalysts that could fix atmospheric carbon and so mitigate climate change. They could discover novel chemical reactions to turn waste plastics into useful raw materials and remove toxic “forever chemicals” from the environment. They could uncover new battery chemistries for safer, more compact energy storage. They could supercharge drug discovery for personalized medicine.</p><p>And that would just be the beginning. We believe quantum-enhanced AI will open up new frontiers in materials science and reshape our ability to understand and manipulate matter at its most fundamental level. Here’s how.</p><h2>How Quantum Computing Will Revolutionize Chemistry</h2><p>To understand how quantum computing and AI could help bend Jacob’s Ladder, it’s useful to look at the classical approximation techniques that are currently used in chemistry. In atoms and molecules, electrons interact with one another in complex ways called electron correlations. These correlations are crucial for accurately describing chemical systems. Many computational methods, such as <a href="https://www.synopsys.com/glossary/what-is-density-functional-theory.html" target="_blank">density functional theory</a> (DFT) or the <a href="https://insilicosci.com/hartree-fock-method-a-simple-explanation/" target="_blank">Hartree-Fock method</a>, simplify these interactions by replacing the intricate correlations with averaged ones, assuming that each electron moves within an average field created by all other electrons. Such approximations work in many cases, but they can’t provide a full description of the system.</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="a woman stirs a white powder inside a glove box." class="rm-shortcode" data-rm-shortcode-id="c0e1bdeb8e874740173f3f02c62eb308" data-rm-shortcode-name="rebelmouse-image" id="40c54" loading="lazy" src="https://spectrum.ieee.org/media-library/a-woman-stirs-a-white-powder-inside-a-glove-box.jpg?id=63745112&width=980"/> </p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="The second shows white powder in test tubes." class="rm-shortcode" data-rm-shortcode-id="5ac7a16946b97de61047d14b9ff28eb7" data-rm-shortcode-name="rebelmouse-image" id="2b1dd" loading="lazy" src="https://spectrum.ieee.org/media-library/the-second-shows-white-powder-in-test-tubes.jpg?id=63745094&width=980"/> </p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="shows a gloved hand holding a silvery disc close to an electronic apparatus." class="rm-shortcode" data-rm-shortcode-id="f3e77cc9b1b4502b2fab5ed6a3cf10f5" data-rm-shortcode-name="rebelmouse-image" id="98787" loading="lazy" src="https://spectrum.ieee.org/media-library/shows-a-gloved-hand-holding-a-silvery-disc-close-to-an-electronic-apparatus.jpg?id=63745089&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">A joint project between Microsoft and Pacific Northwest National Laboratory used AI and high-performance computing to identify potential materials for battery electrolytes. The most promising were synthesized [top and middle] and tested [bottom] at PNNL. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Dan DeLong/Microsoft</small></p><p>Electron correlation is particularly important in systems where the electrons are strongly interacting—as in materials with unusual electronic properties, like high-temperature superconductors—or when there are many possible arrangements of electrons with similar energies—such as compounds containing certain metal atoms that are crucial for catalytic processes.</p><p>In these cases, the simplified approach of DFT or Hartree-Fock breaks down, and more sophisticated methods are needed. As the number of possible electron configurations increases, we quickly reach an “exponential wall” in computational complexity, beyond which classical methods become infeasible.</p><p>Enter the quantum computer. Unlike classical bits, which are either on or off, qubits can exist in superpositions—effectively coexisting in multiple states simultaneously. This should allow them to represent many electron configurations at once, mirroring the complex quantum behavior of correlated electrons. Because quantum computers operate on the same principles as the electron systems they will simulate, they will be able to accurately simulate even strongly correlated systems—where electrons are so interdependent that their behavior must be calculated collectively.</p><h2>AI’s Role in Advancing Computational Chemistry</h2><p>At present, even the computationally cheap methods at the bottom of Jacob’s Ladder are slow, and the ones higher up the ladder are slower still. AI models have emerged as powerful accelerators to such calculations because they can serve as emulators that predict simulation outcomes without running the full calculations. The models can speed up the time it takes to solve problems up and down the ladder by orders of magnitude.</p><p>This acceleration opens up entirely new scales of scientific exploration. In 2023 and 2024, we collaborated with researchers at <a href="https://www.pnnl.gov/" target="_blank">Pacific Northwest National Laboratory</a> (PNNL) on using <a href="https://arxiv.org/abs/2401.04070" rel="noopener noreferrer" target="_blank">advanced AI models</a> to evaluate over 32 million potential battery materials, looking for safer, cheaper, and more environmentally friendly options. This enormous pool of candidates would have taken about 20 years to explore using traditional methods. And yet, within less than a week, <a href="https://spectrum.ieee.org/ai-battery-material" target="_blank">that list was narrowed</a> to 500,000 stable materials and then to 800 highly promising candidates. Throughout the evaluation, the AI models replaced expensive and time-consuming quantum chemistry calculations, in some cases delivering insights half a million times as fast as would otherwise have been the case.</p><p>We then used high-performance computing (HPC) to validate the most promising materials with DFT and AI-accelerated molecular dynamics simulations. The PNNL team then spent about nine months synthesizing and testing one of the candidates—a solid-state electrolyte that uses sodium, which is cheap and abundant, and some other materials, with 70 percent less lithium than conventional lithium-ion designs. The team then built a prototype solid-state battery that they tested over a range of temperatures.</p><p>This potential battery breakthrough isn’t unique. AI models have also dramatically accelerated research in <a href="https://science.nasa.gov/earth/ai-open-science-climate-change/" rel="noopener noreferrer" target="_blank">climate science</a>, <a href="https://www.sciencedirect.com/science/article/pii/S3050585225000217" rel="noopener noreferrer" target="_blank">fluid dynamics</a>, <a href="https://www.simonsfoundation.org/2024/08/26/astrophysicists-use-ai-to-precisely-calculate-universes-settings/" rel="noopener noreferrer" target="_blank">astrophysics</a>, <a href="https://www.nature.com/articles/s44222-025-00349-8" rel="noopener noreferrer" target="_blank">protein design</a>, and <a href="https://www.nature.com/articles/d41586-025-00602-5" rel="noopener noreferrer" target="_blank">chemical and biological discovery</a>. By replacing traditional simulations that can take days or weeks to run, AI is reshaping the pace and scope of scientific research across disciplines.</p><p>However, these AI models are only as good as the quality and diversity of their training data. Whether sourced from high-fidelity simulations or carefully curated experimental results, these data must accurately represent the underlying physical phenomena to ensure reliable predictions. Poor or biased data can lead to misleading outcomes. By contrast, high-quality, diverse datasets—such as those full-accuracy quantum simulations—enable models to generalize across systems and uncover new scientific insights. This is the promise of using quantum computing for training AI models.</p><h2>How to Accelerate Chemical Discovery</h2><p>The real breakthrough will come from strategically combining quantum computing’s and AI’s unique strengths. AI already excels at learning patterns and making rapid predictions. Quantum computers, which are still being scaled up to be practically useful, will excel at capturing electron correlations that classical computers can only approximate. So if you train classical models on quantum-generated data, you’ll get the best of both worlds: the accuracy of quantum delivered at the speed of AI.</p><p>As we learned from the Microsoft-PNNL collaboration on electrolytes, AI models alone can greatly speed up chemical discovery. In the future, quantum-accurate AI models will tackle even bigger challenges. Consider the basic discovery process, which we can think of as a funnel. Scientists begin with a vast pool of candidate molecules or materials at the wide-mouthed top, narrowing them down using filters based on desired properties—such as boiling point, conductivity, viscosity, or reactivity. Crucially, the effectiveness of this screening process depends heavily on the accuracy of the models used to predict these properties. Inaccurate predictions can create a “leaky” funnel, where promising candidates are mistakenly discarded or poor ones are mistakenly advanced.</p><p>Quantum-accurate AI models will dramatically improve the precision of chemical-property predictions. They’ll be able to help identify “first-time right” candidates, sending only the most promising molecules to the lab for synthesis and testing—which will save both time and cost.</p><p>Another key aspect of the discovery process is understanding the chemical reactions that govern how new substances are formed and behave. Think of these reactions as a network of roads winding through a mountainous landscape, where each road represents a possible reaction step, from starting materials to final products. The outcome of a reaction depends on how quickly it travels down each path, which in turn is determined by the energy barriers along the way—like mountain passes that must be crossed. To find the most efficient route, we need accurate calculations of these barrier heights, so that we can identify the lowest passes and chart the fastest path through the reaction landscape.</p><p>Even small errors in estimating these barriers can lead to incorrect predictions about which products will form. Case in point: A slight miscalculation in the energy barrier of an environmental reaction could mean the difference between labeling a compound a “forever chemical” or one that safely degrades over time.</p><p class="shortcode-media shortcode-media-youtube"> <span class="rm-shortcode" data-rm-shortcode-id="70e0b9b540bc0e061b38252e88243293" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/X1aWMYukuUk?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span> </p><p>Accurate modeling of reaction rates is also essential for designing catalysts—substances that speed up and steer reactions in desired directions. Catalysts are crucial in industrial chemical production, carbon capture, and biological processes, among many other things. Here, too, quantum-accurate AI models can play a transformative role by providing the high-fidelity data needed to predict reaction outcomes and design better catalysts.</p><p>Once trained, these AI models, powered by quantum-accurate data, will revolutionize computational chemistry by delivering quantum-level precision. And once the AI models, which run on classical computers, are trained with quantum computing data, researchers will be able to run high-accuracy simulations on laptops or desktop computers, rather than relying on massive supercomputers or future quantum hardware. By making advanced chemical modeling more accessible, these tools will democratize discovery and empower a broader community of scientists to tackle some of the most pressing challenges in health, energy, and sustainability.</p><h2>Remaining Challenges for AI and Quantum Computing</h2><p>By now, you’re probably wondering: When will this transformative future arrive? It’s true that<strong> </strong>quantum computers still struggle with <a href="https://spectrum.ieee.org/quantum-error-correction" target="_blank">error rates</a> and limited lifetimes of usable qubits. And they still need to scale to the size required for meaningful chemistry simulations. Meaningful chemistry simulations beyond the reach of classical computation will require hundreds to thousands of high-quality qubits with error rates of around 10<span><sup>-15</sup></span>, or one error in a quadrillion operations. Achieving this level of reliability will require fault tolerance through redundant encoding of quantum information in logical qubits, each consisting of hundreds of physical qubits, thus requiring a total of about a million physical qubits. Current AI models for chemical-property predictions may not have to be fully redesigned. We expect that it will be sufficient to start with models pretrained on classical data and then fine-tune them with a few results from quantum computers.</p><p> Despite some open questions, the potential rewards in terms of scientific understanding and technological breakthroughs make our proposal a compelling direction for the field. The quantum computing industry has begun to move beyond the early noisy prototypes, and high-fidelity quantum computers with low error rates could be possible <a href="https://www.darpa.mil/research/programs/quantum-benchmarking-initiative" target="_blank">within a decade</a>.</p><p>Realizing the full potential of quantum-enhanced AI for chemical discovery will require focused collaboration between chemists and materials scientists who understand the target problems, experts in quantum computing who are building the hardware, and AI researchers who are developing the algorithms. Done right, quantum-enhanced AI could start to tackle the world’s toughest challenges—from climate change to disease—years ahead of anyone’s expectations. <span class="ieee-end-mark"></span></p>]]></description><pubDate>Mon, 02 Mar 2026 13:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/quantum-chemistry</guid><category>Quantum-computing</category><category>Quantum-chemistry</category><category>Drug-discovery</category><category>Batteries</category><category>Ai-models</category><category>Microsoft</category><dc:creator>Chi Chen</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/illustration-of-a-human-head-in-profile-with-a-spiral-upon-which-human-figures-are-walking-overlaid-on-an-image-of-an-atom.png?id=63744636&amp;width=980"></media:content></item><item><title>Letting Machines Decide What Matters</title><link>https://spectrum.ieee.org/ai-new-physics</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/large-particle-detector-with-circular-structure-person-standing-below.png?id=65005476&width=2000&height=1500&coordinates=26%2C0%2C27%2C0"/><br/><br/><p>In the time it takes you to read this sentence, the <a href="https://spectrum.ieee.org/tag/large-hadron-collider" target="_blank">Large Hadron Collider</a> (LHC) will have smashed billions of particles together. In all likelihood, it will have found exactly what it found yesterday: more evidence to support the <a href="https://home.cern/science/physics/standard-model" rel="noopener noreferrer" target="_blank">Standard Model</a> of particle physics.</p><p>For the engineers who built this 27-kilometer-long ring, this consistency is a triumph. But for theoretical physicists, it has been rather frustrating. As <a href="https://spectrum.ieee.org/u/matthew-hutson" rel="noopener noreferrer" target="_blank">Matthew Hutson</a> reports in “<a data-linked-post="2675068613" href="https://spectrum.ieee.org/particle-physics-ai" target="_blank">AI Hunts for the Next Big Thing in Physics</a>,” the field is currently gripped by a quiet crisis. In an email discussing his reporting, Hutson explains that the Standard Model, which describes the known elementary particles and forces, is not a complete picture. “So theorists have proposed new ideas, and experimentalists have built giant facilities to test them, but despite the gobs of data, there have been no big breakthroughs,” Hutson says. “There are key components of reality we’re completely missing.”</p><p>That’s why researchers are turning artificial intelligence loose on particle physics. They aren’t simply asking AI to comb through accelerator data to confirm existing theories, Hutson explains. They’re asking AI to point the way toward theories that they’ve never imagined. “Instead of looking to support theories that humans have generated,” he says, “unsupervised AI can highlight anything out of the ordinary, expanding our reach into unknown unknowns.” By asking AI to flag anomalies in the data, researchers hope to find their way to “<a href="https://en.wikipedia.org/wiki/Physics_beyond_the_Standard_Model" target="_blank">new physics</a>” that extends the Standard Model. </p><p>On the surface, this article might sound like another “AI for <em><em>X</em></em>” story. As <em><em>IEEE</em></em> <em><em>Spectrum</em></em>’s AI editor, I get a steady stream of pitches for such stories: AI for drug discovery, AI for farming, AI for wildlife tracking. Often what that really means is faster data processing or automation around the edges. Useful, sure, but incremental.</p><p>What struck me in Hutson’s reporting is that this effort feels different. Instead of analyzing experimental data after the fact, the AI essentially becomes part of the instrument, scanning for subtle patterns and deciding in real time what’s interesting. At the LHC, detectors record 40 million collisions per second. There’s simply no way to preserve all that data, so engineers have always had to build filters to decide which events get saved for analysis and which are discarded; nearly everything is thrown away. </p><p>Now those split-second decisions are increasingly handed to machine learning systems running on field-programmable gate arrays (FPGAs) connected to the detectors. The code must run on the chip’s limited logic and memory, and compressing a neural network into that hardware isn’t easy. Hutson describes one theorist pleading with an engineer, “Which of my algorithms fits on your bloody FPGA?”</p><p>This moment is part of a much older pattern. As Hutson writes in the article, new instruments have opened doors to the unexpected throughout the history of science. Galileo’s telescope <a href="https://www.nasa.gov/general/415-years-ago-astronomer-galileo-discovers-jupiters-moons/" target="_blank">revealed moons circling Jupiter</a>. Early microscopes exposed entire worlds of “<a href="https://hekint.org/2018/10/23/van-leeuwenhoeks-discovery-of-animalcules/" target="_blank">animalcules</a>” swimming around. These better tools didn’t just answer existing questions; they made it possible to ask new ones.</p><p>If there’s a crisis in particle physics, in other words, it may not just be about missing particles. It’s about how to look beyond the limits of the human imagination. Hutson’s story suggests that AI might not solve the mysteries of the universe outright, but it could change how we search for answers.</p>]]></description><pubDate>Sun, 01 Mar 2026 11:00:03 +0000</pubDate><guid>https://spectrum.ieee.org/ai-new-physics</guid><category>Large-hadron-collider</category><category>Lhc</category><category>Particle-physics</category><category>Fpga</category><category>Machine-learning</category><dc:creator>Eliza Strickland</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/large-particle-detector-with-circular-structure-person-standing-below.png?id=65005476&amp;width=980"></media:content></item><item><title>A Shapeshifting Supercomputer May Be More Energy Efficient</title><link>https://spectrum.ieee.org/reconfigurable-supercomputer</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/sandia-national-laboratories-supercomputer-spectra.jpg?id=65068162&width=2000&height=1500&coordinates=166%2C0%2C167%2C0"/><br/><br/><p>Late last year, Sandia National Laboratories started testing an unusual type of supercomputer. Unlike conventional <a data-linked-post="2658957928" href="https://spectrum.ieee.org/europe-s-exascale-supercomputer" target="_blank">supercomputers</a>, which consist of large interconnected clusters of CPUs and GPUs, this machine incorporates reconfigurable accelerators that optimize their operations for the particular computation that’s being run. This new architecture, which is similar to field-programmable gate arrays (<a data-linked-post="2650233096" href="https://spectrum.ieee.org/painless-fpga-programming" target="_blank">FPGAs</a>), is built by startup NextSilicon. A key benefit of the approach is that it doesn’t require a software rewrite: the hardware optimizes itself for the software, not vice versa. </p><p>Spectra, which incorporates 128 <a href="https://www.nextsilicon.com/" rel="noopener noreferrer" target="_blank">NextSilicon</a> Maverick-2 accelerators, is still in the investigative phase, says program leader and Sandia senior scientist <a href="https://www.sandia.gov/ccr/staff/james-h-laros/" rel="noopener noreferrer" target="_blank">James Laros</a>. NextSilicon, which has headquarters in Tel Aviv and Minneapolis, claims its accelerators generally use half as much power as Nvidia’s Blackwell while offering a quadruple speed advantage. The power and speed vary depending on the particular application.</p><h2>The Power of Flexibility</h2><p>NextSilicon CEO Elad Raz says typical architectures predict the next instruction then fetch and cache data. “What if you can remove all that overhead?” he wondered. “A lot of people are trying to build a new CPU or a better GPU. Other companies have a software solution,” says Raz. “I wanted to build something with software and hardware collaborating together.”</p><p>The company’s Maverick-2 first runs the application on a CPU and identifies which operations run most frequently. Then, it reconfigures the chip to schedule its work in a way that optimizes data flow. Instead of back-and-forth fetching of data, he says, “you can generate a pipeline.” </p><p>And a key advantage of the company’s design is that users do not have to rewrite their software in order to run it more efficiently on the system. The hardware adapts to the software, not vice versa.</p><p>Most of the applications Sandia runs are constrained by memory bandwidth, says Laros. “What if we can go faster because we don’t have to go back to the main memory?” That’s the potential of the Spectra architecture.</p><p><span>Raz says Maverick uses half as much power as Nvidia’s Blackwell and can perform HPCG, a supercomputing benchmark, twice as fast; it performs PageRank, another benchmark, 10 times as fast. Sandia scientists are currently assessing Spectra’s performance when running molecular dynamics simulations, which predict the movements of atoms and are widely used in physics and materials science, and other core codes used by the U.S. Department of Energy. “Where it will provide a benefit is if we can get better performance for types of apps that don’t run well on GPUs, or if we can get the same performance with better energy efficiency,” Laros says.</span></p><h2>Supporting the Mission Through Experimentation</h2><p>Sandia performs computer simulations to maintain the United States’ nuclear arsenal. “We’ve replaced testing with simulation and computing,” Laros says. Because of the high stakes of this mission, the lab has to “make sure we’re not putting all our eggs in one basket,” he says. If a company whose technology the U.S. government relies on for nuclear stockpile stewardship should go out of business, the government needs to have alternatives. “We maintain a pipeline of overlapping technologies,” Laros says.</p><p>Spectra is part of Sandia’s <a href="https://vanguard.sandia.gov/" target="_blank">Vanguard</a> program, which allows the government to partner with startups to test out and help develop early-stage high-performance computing technologies. “The goal is to test them for our advanced simulation and computing mission codes,” Laros says.</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="Close-up several wires plugged into the back of a server rack." class="rm-shortcode" data-rm-shortcode-id="b59df3bc1edf177d617ec79d9a0a2e49" data-rm-shortcode-name="rebelmouse-image" id="99b63" loading="lazy" src="https://spectrum.ieee.org/media-library/close-up-several-wires-plugged-into-the-back-of-a-server-rack.jpg?id=65068423&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Penguin Solutions integrated the thermal-management and power-distribution systems for Spectra, and led the installation at Sandia National Laboratories. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Craig Fritz/Sandia National Laboratories</small></p><p>Sandia runs a large portion of its code on CPUs, says Laros. They’ve adopted GPU based systems built by Nvidia as well. These systems offer speed advantages, but they require lab staff to port their code. “It took us hundreds of hours,” says Laros. And there are important scientific simulations that don’t run well on GPUs, including Monte Carlo methods, which can be used to assess complex risks.</p><p>Laros says it’s unusual right now to find a computing startup focusing on high-performance scientific computing—“It’s all about AI” these days, he says. Next Silicon is developing hardware that the company hopes will have advantages for both, thanks to its promised power savings. Power availability is a major constraint on large-scale AI data centers today. Raz hopes NextSilicon’s accelerators will offer an advantage by enabling more efficient performance for a given amount of electricity consumption.</p> The Vanguard program allows the government to test the potential of risky technologies. “You’re going to fail once in a while,” says Laros. “Our goal is to do very advanced technology discovery. We prove it out. Other labs and other commercial industries will follow.”]]></description><pubDate>Fri, 27 Feb 2026 14:23:28 +0000</pubDate><guid>https://spectrum.ieee.org/reconfigurable-supercomputer</guid><category>Computing</category><category>Sandia-national-laboratories</category><category>Energy-efficiency</category><category>Supercomputing</category><dc:creator>Katherine Bourzac</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/sandia-national-laboratories-supercomputer-spectra.jpg?id=65068162&amp;width=980"></media:content></item><item><title>AI Is Acing Math Exams Faster Than Scientists Write Them</title><link>https://spectrum.ieee.org/ai-math-benchmarks</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/line-graph-demonstrates-how-google-deepminds-aletheia-ai-scores-at-least-5-percent-higher-on-ph-d-math-exercises-than-the-lat.jpg?id=65007034&width=2000&height=1500&coordinates=166%2C0%2C167%2C0"/><br/><br/><p><span>Mathematics is often regarded as the ideal domain for measuring AI progress effectively. Math’s step-by-step logic is easy to track, and its definitive, automatically verifiable answers remove any human or subjective factors. But AI systems are improving at such a pace that math </span><a href="https://spectrum.ieee.org/melanie-mitchell" target="_self">benchmarks are struggling to keep up</a><span>.</span></p><p>Way back in November 2024, nonprofit research organization Epoch AI quietly released <a href="https://doi.org/10.48550/arXiv.2411.04872" target="_blank">FrontierMath</a>. A standardized, rigorous benchmark, FrontierMath was designed to measure the mathematical reasoning capabilities of the latest AI tools.</p><p>“It’s a bunch of really hard math problems,” explains <a href="https://epoch.ai/team" target="_blank">Greg Burnham</a>, Epoch AI senior researcher. “Originally, it was 300 problems that we now call tiers 1–3, but having seen AI capabilities really speed up, there was a feeling that we had to run to stay ahead, so now there’s a special challenge set of extra carefully constructed problems that we call tier 4.”</p><p>To a rough approximation, tiers 1–4 go from advanced undergraduate through to early postdoc-level mathematics. When introduced, state-of-the-art AI models were unable to solve more than 2 percent of the problems FrontierMath contained. <a href="https://epoch.ai/frontiermath/tiers-1-4" target="_blank">Fast forward to today</a>: The best publicly available AI models, such as GPT-5.2 and Claude Opus 4.6, are solving over 40 percent of FrontierMath’s 300 tier 1–3 problems, and over 30 percent of the 50 tier 4 problems.</p><h2>AI takes on Ph.D.-level mathematics</h2><p>And this dizzying pace of advancement is showing no signs of abating. For example, just recently <a href="https://deepmind.google/blog/accelerating-mathematical-and-scientific-discovery-with-gemini-deep-think/" target="_blank">Google DeepMind announced</a> that Aletheia, an experimental AI system derived from Gemini Deep Think, <a href="https://doi.org/10.48550/arXiv.2601.23245" target="_blank">achieved publishable Ph.D.-level research results</a>. Though obscure mathematically—it was calculated with certain structure constants in arithmetic geometry called eigenweights—the result is significant in terms of AI development.</p><p>“They’re claiming it was essentially autonomous, meaning a human wasn’t guiding the work, and it’s publishable,” Burnham says. “It’s definitely at the lower end of the spectrum of work that would get a mathematician excited, but it’s new—it’s something we truly haven’t really seen before.”</p><p>To place this achievement in context, every FrontierMath problem has a known answer that a human has derived. Though a human could probably have achieved Aletheia’s result “if they sat down and steeled themselves for a week,” says Burnham, no human had ever done so.</p><p>Aletheia’s results and other recent achievements by AI mathematicians point to new, tougher benchmarks being needed to understand AI capabilities—and fast, because existing ones will soon become irrelevant. “There are easier math benchmarks that are already obsolete, several generations of them,” says Burnham. “FrontierMath will probably saturate [Ed. note: This means that state-of-the-art AI models score 100 percent] within the next two years—could be faster.”</p><h2>The First Proof challenge</h2><p>To begin to address this problem, on 6 February, a group of 11 highly distinguished mathematicians <a href="https://doi.org/10.48550/arXiv.2602.05192" rel="noopener noreferrer" target="_blank">proposed the First Proof challenge</a>, a set of 10 extremely difficult math questions that arose naturally in the authors’ research processes, and whose proofs are roughly five pages or less and had not been shared with anyone. <a href="https://1stproof.org/" rel="noopener noreferrer" target="_blank">The First Proof challenge</a> was a preliminary effort to assess the capabilities of AI systems in solving research-level math questions on their own.</p><p>Generating serious buzz in the math community, professional and amateur mathematicians, and teams including OpenAI, all stepped up to the challenge. But by the time the authors <a href="https://codeberg.org/tgkolda/1stproof/src/branch/main/2026-02-batch/FirstProofSolutionsComments.pdf" rel="noopener noreferrer" target="_blank">posted the proofs</a> on 14 February, no one had submitted correct solutions to all 10 problems.</p><p>In fact, far from it. The authors themselves only solved two of the 10 problems using Gemini 3.0 Deep Think and ChatGPT 5.2 Pro. And most outside submissions fared little better, apart from OpenAI and a small Aletheia team at Google DeepMind. With “limited human supervision,” OpenAI’s most advanced internal AI system <a href="https://openai.com/index/first-proof-submissions/" rel="noopener noreferrer" target="_blank">solved five of the 10 problems</a>, with Aletheia achieving similar outcomes—results met with a spectrum of emotions by different members of the mathematics community, from awe to disappointment. The team behind First Proof plans an even tougher <a href="https://1stproof.org/" rel="noopener noreferrer" target="_blank">second round on 14 March</a>.</p><h2>A new frontier for AI</h2><p>“I think First Proof is terrific: It’s as close as you could realistically get to putting an AI system in the shoes of a mathematician,” says Burnham. Though he admires how First Proof tests AI’s mathematical utility for a wide range of mathematics and mathematicians, Epoch AI has its own new approach to testing—<a href="https://epoch.ai/frontiermath/open-problems" rel="noopener noreferrer" target="_blank">FrontierMath: Open Problems</a>. Uniquely, the pilot benchmark consists of 16 open problems (with more to follow) from research mathematics that professional mathematicians have tried and failed to solve. Since Open Problems’ <a href="https://epochai.substack.com/p/introducing-frontiermath-open-problems" rel="noopener noreferrer" target="_blank">release on 27 January</a>, none have been solved by an AI.</p><p>“With Open Problems, we’ve tried to make it more challenging,” says Burnham. “The baseline on its own would be publishable, at least in a specialty journal.” What’s more, each question is designed so that it can be automatically graded. “This is a bit counterintuitive,” Burnham adds. “No one knows the answers, but we have a computer program that will be able to judge whether the answer is right or not.”</p><p>Burnham sees First Proof and Open Problems as being complementary. “I would say understanding AI capabilities is a more-the-merrier situation,” he adds. “AI has gotten to the point where it’s—in some ways—better than most Ph.D. students, so we need to pose problems where the answer would be at least moderately interesting to some human mathematicians, not because AI was doing it but because it’s mathematics that human mathematicians care about.”</p>]]></description><pubDate>Wed, 25 Feb 2026 16:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/ai-math-benchmarks</guid><category>Ai-benchmarks</category><category>Mathematics</category><category>Large-language-models</category><category>Artificial-intelligence</category><dc:creator>Benjamin Skuse</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/line-graph-demonstrates-how-google-deepminds-aletheia-ai-scores-at-least-5-percent-higher-on-ph-d-math-exercises-than-the-lat.jpg?id=65007034&amp;width=980"></media:content></item><item><title>AI’s Math Tricks Don’t Work for Scientific Computing</title><link>https://spectrum.ieee.org/number-formats-ai-scientific-computing</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/abstract-image-with-colorful-bar-graphs-numbers-and-lines-on-a-pastel-background.png?id=64959805&width=2000&height=1500&coordinates=50%2C0%2C51%2C0"/><br/><br/><p>AI has driven an explosion of new number formats—the ways in which numbers are represented digitally. Engineers are looking at every possible way <a data-linked-post="2671036628" href="https://spectrum.ieee.org/mems-time" target="_blank">to save computation time and energy</a>, including shortening the number of bits used to represent data. But what works for AI doesn’t necessarily work for scientific computing, be it for computational physics, biology, fluid dynamics, or engineering simulations. <em><em>IEEE Spectrum</em></em> spoke with <a href="https://laslo.hunhold.de/" target="_blank">Laslo Hunhold</a>, who recently joined Barcelona-based <a href="https://openchip.com/" target="_blank">Openchip</a> as an AI engineer, about his efforts to develop a bespoke number format for scientific computing.</p><h3>LASLO HUNHOLD</h3><br/><p><a href="https://laslo.hunhold.de/" rel="noopener noreferrer" target="_blank">Laslo Hunhold</a> is a senior AI accelerator engineer at Barcelona-based startup Openchip. He recently completed a Ph.D. in computer science from the University of Cologne, in Germany.</p><p><strong>What makes number formats interesting to you?</strong></p><p><strong>Laslo Hunhold: </strong>I don’t know another example of a field that so few are interested in but has such a high impact. If you make a number format that’s 10 percent more [energy] efficient, it can translate to all applications being 10 percent more efficient, and you can save a lot of energy.</p><p><strong>Why are there so many new number formats?</strong></p><p><strong>Hunhold: </strong>For decades, computer users had it really easy. They could just buy new systems every few years, and they would have performance benefits for free. But this hasn’t been the case for the last 10 years. In computers, you have a certain number of bits used to represent a single number, and for years the default was 64 bits. And for AI, companies noticed that they don’t need 64 bits for each number. So they had a strong incentive to go down to 16, 8, or even 2 bits [to save energy]. The problem is, the dominating standard for representing numbers in 64 bits is not well designed for lower bit counts. So in the AI field, they came up with new formats which are more tailored toward AI.</p><p><strong>Why does AI need different number formats than scientific computing? </strong></p><p><strong>Hunhold: </strong>Scientific computing needs high dynamic range: You need very large numbers, or very small numbers, and very high accuracy in both cases. The 64-bit standard has an excessive dynamic range, and it is many more bits than you need most of the time. It’s different with AI. The numbers usually follow a specific distribution, and you don’t need as much accuracy.</p><p><strong>What makes a number format “good”?</strong></p><p><strong>Hunhold: </strong>You have infinite numbers but only finite bit representations. So you need to decide how you assign numbers. The most important part is to represent numbers that you’re actually going to use. Because if you represent a number that you don’t use, you’ve wasted a representation. The simplest thing to look at is the dynamic range. The next is distribution: How do you assign your bits to certain values? Do you have a uniform distribution, or something else? There are infinite possibilities.</p><p><strong>What motivated you to introduce the takum number format?</strong></p><p><strong>Hunhold: </strong><span>Takums are based on <a href="https://spectrum.ieee.org/floating-point-numbers-posits-processor" target="_blank">posits</a>. In posits, the numbers that get used more frequently can be represented with more density. But posits don’t work for scientific computing, and this is a huge issue. They have a high density for [numbers close to one], which is great for AI, but the density falls off sharply once you look at larger or smaller values. People have been proposing dozens of number formats in the last few years, but takums are the only number format that’s actually tailored for scientific computing. I found the dynamic range of values you use in scientific computations, if you look at all the fields, and designed takums such that when you take away bits, you don’t reduce that dynamic range</span></p><p><em>This article appears in the March 2026 print issue as “Laslo Hunhold.”</em></p>]]></description><pubDate>Mon, 23 Feb 2026 13:00:03 +0000</pubDate><guid>https://spectrum.ieee.org/number-formats-ai-scientific-computing</guid><category>Typedepartments</category><category>5-questions</category><category>Artificial-intelligence</category><category>Computing</category><category>Scientific-computing</category><dc:creator>Dina Genkina</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/abstract-image-with-colorful-bar-graphs-numbers-and-lines-on-a-pastel-background.png?id=64959805&amp;width=980"></media:content></item><item><title>Laser-Written Glass Could Store Data for Millennia</title><link>https://spectrum.ieee.org/glass-data-storage-microsoft-silica</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-square-labelled-microsoft-azure-with-7-vertical-segments-each-a-slightly-different-shade-of-blue-with-a-great-deal-of-textur.jpg?id=64953537&width=2000&height=1500&coordinates=166%2C0%2C167%2C0"/><br/><br/><p><span>Huge amounts of data could be archived in a compact format for millennia, by using lasers to encode that data into glass.</span></p><p>In 2028, the amount of data that the world is projected to generate—in the form of photos, videos, emails, documents, and other files—could reach <a href="https://blog.westerndigital.com/giving-hdd-rare-earth-elements-new-life/" target="_blank">about 394 zettabytes</a> (394 trillion gigabytes), according to the market analyst firm <a href="https://www.idc.com/" target="_blank">International Data Corporation</a>. That’s a nearly threefold increase from 2023.This demand has driven research toward <a href="https://spectrum.ieee.org/dna-data-storage" target="_self">radically new ways of storing data</a>, in part because all mainstream electronic data storage techniques break down over time. For example, most digital archives are currently saved on <a href="https://spectrum.ieee.org/why-the-future-of-data-storage-is-still-magnetic-tape" target="_self">magnetic tape</a>, which only lasts about <a href="https://www.clir.org/2022/08/study-reveals-new-findings-on-longevity-of-legacy-magnetic-audio-tape/" target="_blank">30 years</a>.</p><p>One possibility that scientists have explored is storing data in glass because of its resistance to moisture, temperature fluctuations, and electromagnetic interference. However, previous attempts exploring this strategy faced problems with data density, writing throughput, and energy efficiency.</p><p>In a study, Microsoft scientists sought to overcome those deficiencies by using femtosecond lasers, which fire high-power laser pulses just quadrillionths of a second long. The researchers focused a laser beam on a point within glass to modify its optical properties and encode data as voxels, the 3D equivalent of pixels. The researchers published their <a href="https://www.nature.com/articles/s41586-025-10042-w" target="_blank">findings</a> online on 18 February in <em>Nature</em>.</p><h2>Advanced Glass Data-Storage Technology</h2><p>Each voxel the scientists encoded was about 0.5 micrometers large in each direction and separated from each other by about 6 µm. They could then read out the data using an automated microscope. Machine learning models used during reading help account for the effects of random errors and optical interference between voxels.</p><p>The new archival system, called <a href="https://www.microsoft.com/en-us/research/project/project-silica/" target="_blank">Silica</a>, can store up to 1.59 gigabits per cubic millimeter. This amounts to 4.84 terabytes stored in roughly 300 layers within a 12-square-centimeter, 2-millimeter-deep glass chip, the equivalent of about 2 million printed books or 5,000 ultrahigh-definition 4K films.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Close-up of optical equipment in a lab" class="rm-shortcode" data-rm-shortcode-id="4ede2ecf100600aa459496c818b567f5" data-rm-shortcode-name="rebelmouse-image" id="73177" loading="lazy" src="https://spectrum.ieee.org/media-library/close-up-of-optical-equipment-in-a-lab.jpg?id=64953552&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Femtosecond laser pulses can write data into glass in the form of tiny 3D voxels.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Microsoft Research</small></p><p>One drawback of previous attempts at glass data storage was how they often relied on multiple laser pulses to write each voxel. This led to reduced throughput and high power demands. <span>To overcome this challenge, the scientists developed two different types of voxels, called phase voxels and birefringent voxels, which each required only a single laser pulse to write. <strong></strong></span></p><p>“This significantly reduces the power required from the laser to store data, and it does not require the laser focus to alternate between staying in the same place to deliver multiple pulses and movement to the next location,” says <a href="https://www.microsoft.com/en-us/research/people/rjblack/" target="_blank">Richard Black</a>, the research director for Microsoft’s Project Silica. “Consequently, the laser focus can be swept rapidly across the glass, enabling writing speed only limited by the speed of the femtosecond laser itself.” All in all, by firing millions of laser pulses per second, Silica could write up to 25.6 megabits per second with a single beam.</p><p>Both of these kinds of voxels modify the speed at which light travels through the glass, which is based off a factor called the refractive index. A phase voxel alters the refractive index uniformly in all directions, whereas a birefringent voxel changes the refractive index in a direction-dependent manner.</p><p>With birefringent voxels, the scientists could achieve significantly better data density, write throughput, and energy efficiency. However, these voxels required high-purity silica glasses. In comparison, phase voxels can potentially be written in any durable transparent material, such as borosilicate glass, which is widely available in cookware. Phase voxels also require simpler writing and reading hardware.</p><p>To assess the long-term stability of Silica, the scientists repeatedly heated inscribed glass plates up to 500 °C, simulating the long-term aging of the glass at lower temperatures. Their results suggest the data on the plates might be readable for more than 10,000 years at 290 °C, and likely even longer at ambient temperature, an endurance far exceeding conventional electronic data-storage techniques.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Laboratory equipment with optical components and instruments on a metal table." class="rm-shortcode" data-rm-shortcode-id="3970475db130e3ffa67f1dc256560226" data-rm-shortcode-name="rebelmouse-image" id="146f7" loading="lazy" src="https://spectrum.ieee.org/media-library/laboratory-equipment-with-optical-components-and-instruments-on-a-metal-table.jpg?id=64953558&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">An automated microscope can read out the voxels carved into glass to recover the stored data.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Microsoft research</small></p><p>“This is not a replacement for everyday storage like [solid-state drives] or hard drives,” Black says. “It’s designed for data you want to write once and preserve for a very long time.”</p><p>The clearest applications “are archival—anywhere data must survive for centuries such as national libraries, scientific data, or cultural records,” Black says. “It’s also compelling for cloud archives, where data is written once and kept indefinitely.”</p><p>A <a href="https://ceramics.org/ceramic-tech-today/glass-viscosity-calculations-definitively-debunk-the-myth-of-observable-flow-in-medieval-windows/" target="_blank">common misconception about glass</a> is that it flows over a few centuries. Black notes that “at room temperature, glass is effectively a solid and does not flow on any meaningful time scale.” At normal storage conditions, “the glass and the data structures within it are essentially unchanged for far longer than human history.”</p><p>Black cautions that questions remain about how feasible this strategy can prove at scale, as femtosecond lasers are currently expensive. The researchers hope that by sharing their results, other groups can build on their work to help make the technology accessible.</p><p><em>This article appears in the April 2026 print issue as “Microsoft Boosts Density of Forever Storage in Glass.”</em></p>]]></description><pubDate>Wed, 18 Feb 2026 16:00:03 +0000</pubDate><guid>https://spectrum.ieee.org/glass-data-storage-microsoft-silica</guid><category>Microsoft</category><category>Data-storage</category><category>Lasers</category><dc:creator>Charles Q. Choi</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-square-labelled-microsoft-azure-with-7-vertical-segments-each-a-slightly-different-shade-of-blue-with-a-great-deal-of-textur.jpg?id=64953537&amp;width=980"></media:content></item><item><title>Estimating Surface Heating of an Atmospheric Reentry Vehicle With Simulation</title><link>https://event.on24.com/wcc/r/5204256/9B4BA35D454493C7D9829FE90B5A0ABD</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/comsol-logo.png?id=27157944&width=980"/><br/><br/><p>Join Hannah Alpert (NASA Ames) to explore thermal data from the record-breaking 6-meter LOFTID inflatable aeroshell. Learn how COMSOL Multiphysics® was used to perform inverse analysis on flight thermocouple data, validating heat flux gauges and preflight CFD predictions. Attendees will gain technical insights into improving thermal models for future HIAD missions, making this essential for engineers seeking to advance atmospheric reentry design. The session concludes with a live Q&A.<br/></p><p><span><a href="https://event.on24.com/wcc/r/5204256/9B4BA35D454493C7D9829FE90B5A0ABD" target="_blank">Register now to watch this free on-demand webinar!</a></span></p>]]></description><pubDate>Tue, 17 Feb 2026 19:27:00 +0000</pubDate><guid>https://event.on24.com/wcc/r/5204256/9B4BA35D454493C7D9829FE90B5A0ABD</guid><category>Thermal-management</category><category>Reentry-vehicle</category><category>Multiphysics-simulation</category><category>Thermal-data</category><category>Type-webinar</category><dc:creator>COMSOL</dc:creator><media:content medium="image" type="image/png" url="https://assets.rbl.ms/27157944/origin.png"></media:content></item><item><title>We’re Measuring Data Center Sustainability Wrong</title><link>https://spectrum.ieee.org/data-center-sustainability-metrics</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-pile-of-discarded-old-phones.jpg?id=64070651&width=2000&height=1500&coordinates=416%2C0%2C417%2C0"/><br/><br/><p>In 2024, Google <a href="https://blog.google/company-news/outreach-and-initiatives/sustainability/environmental-report-2025/" rel="noopener noreferrer" target="_blank">claimed</a> that its data centers are 1.5 times more energy-efficient than the industry average. In 2025, Microsoft <a href="https://cdn-dynmedia-1.microsoft.com/is/content/microsoftcorp/microsoft/msc/documents/presentations/CSR/2025-Microsoft-Environmental-Sustainability-Report-PDF.pdf" rel="noopener noreferrer" target="_blank">committed</a> billions to nuclear power for AI workloads. The data center industry tracks power-usage effectiveness to three decimal places and optimizes water usage intensity with machine precision. We report direct emissions and energy emissions with religious fervor.</p><div class="rm-embed embed-media"><iframe height="110px" id="noa-web-audio-player" src="https://embed-player.newsoveraudio.com/v4?key=q5m19e&id=https://spectrum.ieee.org/data-center-sustainability-metrics&bgColor=F5F5F5&color=1b1b1c&playColor=1b1b1c&progressBgColor=F5F5F5&progressBorderColor=bdbbbb&titleColor=1b1b1c&timeColor=1b1b1c&speedColor=1b1b1c&noaLinkColor=556B7D&noaLinkHighlightColor=FF4B00&feedbackButton=true" style="border: none" width="100%"></iframe></div><p><span>These are laudable advances, but these metrics account for only 30 percent of total emissions from the IT sector. The majority of the emissions are not directly from data centers or the energy they use, but from the end-user devices that actually access the data centers, emissions due to manufacturing the hardware, and </span><a href="https://spectrum.ieee.org/green-software" target="_self">software inefficiencies</a><span>. We are frantically optimizing less than a third of the IT sector’s environmental impact, while the bulk of the problem goes unmeasured.</span></p><p>Incomplete regulatory frameworks are part of the problem. In Europe, the Corporate Sustainability Reporting Directive (<a href="https://www.csrdreadiness.com/?utm_term=corporate%20sustainability%20reporting%20directive&utm_campaign=Website+traffic-Search-csrdreadiness-1&utm_source=adwords&utm_medium=ppc&hsa_acc=8049917490&hsa_cam=21799253197&hsa_grp=174334210531&hsa_ad=716504436475&hsa_src=g&hsa_tgt=kwd-1250753155894&hsa_kw=corporate%20sustainability%20reporting%20directive&hsa_mt=p&hsa_net=adwords&hsa_ver=3&gad_source=1&gad_campaignid=21799253197&gbraid=0AAAAAovX5IFUb0B4kiIuBnogJDzencxuq&gclid=Cj0KCQiAhOfLBhCCARIsAJPiopNSX804TFv2FoZyU-WkNfA6MkLIxGnqI0XtmYd6jzOYXXsq4ORQmbIaAv6vEALw_wcB" target="_blank">CSRD</a>) now requires 11,700 companies to report emissions using these incomplete frameworks. The next phase of the directive, covering 40,000+ additional companies, was originally scheduled for 2026 (but is likely delayed to 2028). In the United States, the standards body responsible for IT sustainability metrics (<a href="https://www.iso.org/committee/654019.html" target="_blank">ISO/IEC JTC 1/SC 39</a>) is conducting active revision of its standards through 2026, with a key plenary meeting in May 2026.</p><p>The time to act is now. If we don’t fix the measurement frameworks, we risk locking in incomplete data collection and optimizing a fraction of what matters for the next 5 to 10 years, before the next major standards revision.</p><h2>The limited metrics</h2><p>Walk into any modern data center, and you’ll see sustainability instrumentation everywhere. Power-usage efficiency (PUE) monitors track every watt. Water-usage efficiency (WUE) systems measure water consumption down to the gallon. Sophisticated monitoring captures everything from server utilization to cooling efficiency to renewable energy percentages.</p><p>But here’s what those measurements miss: End-user devices globally emit 1.5 to 2 times more carbon than all data centers combined, according to McKinsey’s 2022 <a href="https://www.mckinsey.com/de/~/media/mckinsey/business%20functions/mckinsey%20digital/our%20insights/the%20green%20it%20revolution%20a%20blueprint%20for%20cios%20to%20combat%20climate%20change/the-green-it-revolution-a-blueprint-for-cios-to-combat-climate-change.pdf" rel="noopener noreferrer" target="_blank">report</a>. The smartphones, laptops, and tablets we use to access those ultra-efficient data centers are the bigger problem.</p><p class="pull-quote">Data center operations, as measured by power-usage efficiency, account for only 24 percent of the total emissions.</p><p>On the conservative end of the range from McKinsey’s report, devices emit 1.5 times as much as data centers. That means that data centers make up 40 percent of total IT emissions, while devices make up 60 percent.</p><p>On top of that, approximately <a href="https://cdn-dynmedia-1.microsoft.com/is/content/microsoftcorp/microsoft/msc/documents/presentations/CSR/2025-Microsoft-Environmental-Sustainability-Report-PDF.pdf" rel="noopener noreferrer" target="_blank">75 percent</a> of device emissions occur not during use, but during manufacturing—this is so-called embodied carbon. For data centers, only 40 percent is embodied carbon, and <a href="https://download.schneider-electric.com/files?p_Doc_Ref=SPD_WP99_EN&p_enDocType=White+Paper&p_File_Name=WP99_V1_EN.pdf" rel="noopener noreferrer" target="_blank">60 percent</a> comes from operations (as measured by PUE).</p><p>Putting this together, data center operations, as measured by PUE, account for only 24 percent of the total emissions. Data center embodied carbon is 16 percent, device embodied carbon is 45 percent, and device operation is 15 percent.</p><p>Under the EU’s current CSRD framework, companies must report their emissions in three categories: direct emissions from owned sources, indirect emissions from purchased energy, and a third category for everything else.</p><p>This “everything else” category does include device emissions and embodied carbon. However, those emissions are reported as aggregate totals broken down by accounting category—capital goods, purchased goods and services, use of sold products—but not by product type. How much comes from end-user devices versus data center infrastructure, or employee laptops versus network equipment, remains murky, and therefore, unoptimized.</p><h2>Embodied carbon and hardware reuse</h2><p>Manufacturing a single smartphone<a href="https://www.sciencedirect.com/science/article/pii/S1364032123002794" rel="noopener noreferrer" target="_blank"> generates</a> approximately 50 kilograms CO<sub>2</sub> equivalent (CO<span><sub>2</sub></span>e). For a laptop, it’s 200 kg CO<span><span><sub>2</sub></span></span>e. With 1 billion smartphones replaced annually, that’s 50 million tonnes of CO<sub>2</sub>e per year just from smartphone manufacturing, before anyone even turns them on.<strong> </strong>On average, smartphones are replaced every two years, laptops every three to four years, and printers every five years. Data center servers are replaced approximately every five years.</p><p class="pull-quote">Extending smartphone life cycles to three years instead of two would reduce annual manufacturing emissions by 33 percent. At scale, this dwarfs data center optimization gains.</p><p>There are programs geared toward reusing old components that are still functional and integrating them into new servers. GreenSKUs and similar initiatives show that 8 percent reductions in embodied carbon are<a href="https://www.microsoft.com/en-us/research/wp-content/uploads/2024/03/2024-GreenSKU-ISCA2024.pdf" target="_blank"> achievable</a>. But these remain pilot programs, not systematic approaches. And critically, they’re measured only in the data center context, not across the entire IT stack.</p><p>Imagine <a href="https://spectrum.ieee.org/carfax-used-pcs" target="_self">applying</a> the same circular economy principles to devices. With over 2 billion laptops in existence globally and two- to three-year replacement cycles, even modest lifespan extensions create massive emission reductions. Extending smartphone life cycles to three years instead of two would reduce annual manufacturing emissions by 33 percent. At scale, this dwarfs data center optimization gains.</p><p>Yet data center reuse gets measured, reported, and optimized. Device reuse doesn’t, because the frameworks don’t require it.</p><h2>The invisible role of software</h2><p>Leading load balancer infrastructure across IBM Cloud, I see how software architecture decisions ripple through energy consumption. Inefficient code doesn’t just slow things down—it drives up both data center power consumption and device battery drain.</p><p>For example, University of Waterloo researchers <a href="https://spectrum.ieee.org/data-center-energy-consumption" target="_self">showed</a> that they can reduce 30 percent of energy use in data centers by changing just 30 lines of code. From my perspective, this result is not an anomaly—it’s typical. Bad software architecture forces unnecessary data transfers, redundant computations, and excessive resource use. But unlike data center efficiency, there’s no commonly accepted metric for software efficiency.</p><p>This matters now more than ever. With AI workloads driving massive data center expansion—projected to consume 6.7 to 12 percent of total U.S. electricity by 2028, <a href="https://eta-publications.lbl.gov/sites/default/files/2024-12/lbnl-2024-united-states-data-center-energy-usage-report_1.pdf" target="_blank">according</a> to Lawrence Berkeley National Laboratory—software efficiency becomes critical.</p><h2>What needs to change</h2><p>The solution isn’t to stop measuring data center efficiency. It’s to measure device sustainability with the same rigor. Specifically, standards bodies (particularly ISO/IEC JTC 1/SC 39 WG4: Holistic Sustainability Metrics) should extend frameworks to include device life-cycle tracking, software efficiency metrics, and hardware reuse standards.</p><p>To track device life cycles, we need standardized reporting of device embodied carbon, broken out separately by device. One aggregate number in an “everything else” category is insufficient. We need specific device categories with manufacturing emissions and replacement cycles visible.</p><p>To include software efficiency, I advocate developing a PUE-equivalent for software, such as energy per transaction, per API call, or per user session. This needs to be a reportable metric under sustainability frameworks so companies can demonstrate software optimization gains.</p><p>To encourage hardware reuse, we need to systematize reuse metrics across the full IT stack—servers and devices. This includes tracking repair rates, developing large-scale refurbishment programs, and tracking component reuse with the same detail currently applied to data center hardware.</p><p>To put it all together, we need a unified IT emission-tracking dashboard. CSRD reporting should show device embodied carbon alongside data center operational emissions, making the full IT sustainability picture visible at a glance.</p><p>These aren’t radical changes—they’re extensions of measurement principles already proven in the data center context. The first step is acknowledging what we’re not measuring. The second is building the frameworks to measure it. And the third is demanding that companies report the complete picture—data centers and devices, servers and smartphones, infrastructure and software.</p><p>Because you can’t fix what you can’t see. And right now, we’re not seeing 70 percent of the problem.</p>]]></description><pubDate>Tue, 17 Feb 2026 15:00:03 +0000</pubDate><guid>https://spectrum.ieee.org/data-center-sustainability-metrics</guid><category>Data-centers</category><category>Sustainability</category><category>Energy</category><category>Software</category><dc:creator>Arjun Sharma</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-pile-of-discarded-old-phones.jpg?id=64070651&amp;width=980"></media:content></item><item><title>This Former Physicist Helps Keep the Internet Secure</title><link>https://spectrum.ieee.org/network-security-engineer-alan-dekok</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/person-in-blue-suit-examines-illuminated-server-racks-in-a-dimly-lit-room.png?id=64440487&width=2000&height=1500&coordinates=110%2C0%2C111%2C0"/><br/><br/><p>When <a href="https://www.linkedin.com/in/alandekok/" rel="noopener noreferrer" target="_blank">Alan DeKok</a> began a side project in network security, he didn’t expect to start a 27-year career. In fact, he didn’t initially set out to work in computing at all.</p><p>DeKok studied nuclear physics before making the switch to a part of network computing that is foundational but—like nuclear physics—largely invisible to those not directly involved in the field. Eventually, a project he started as a hobby became a full-time job: maintaining one of the primary systems that helps keep the internet secure.</p><h3>Alan DeKok</h3><br/><p><strong>Employer</strong></p><p>InkBridge Networks</p><p><strong>Occupation</strong></p><p>CEO</p><p><strong>Education </strong></p><p>Bachelor’s degree in physics, Carleton University; master’s degree in physics, Carleton University</p><p>Today, he leads the <a href="https://www.freeradius.org/" rel="noopener noreferrer" target="_blank">FreeRADIUS Project</a>, which he cofounded in the late 1990s to develop what is now the most widely used Remote Authentication Dial-In User Service (RADIUS) software. FreeRADIUS is an open-source server that provides back-end authentication for most major <a href="https://spectrum.ieee.org/tag/internet-service-providers" target="_blank">internet service providers</a>. It’s used by global financial institutions, Wi-Fi services like <a href="https://eduroam.org/" rel="noopener noreferrer" target="_blank">Eduroam</a>, and Fortune 50 companies. DeKok is also CEO of <a href="https://www.inkbridgenetworks.com/" rel="noopener noreferrer" target="_blank">InkBridge Networks</a>, which maintains the server and provides support for the companies that use it.</p><p>Reflecting on nearly three decades of experience leading FreeRADIUS, DeKok says he became an expert in remote authentication “almost by accident,” and the key to his career has largely been luck. “I really believe that it’s preparing yourself for luck, being open to it, and having the skills to capitalize on it.”</p><h2>From Farming to Physics</h2><p>DeKok grew up on a farm outside of Ottawa growing strawberries and raspberries. “Sitting on a tractor in the heat is not particularly interesting,” says DeKok, who was more interested in working with 8-bit computers than crops. As a student at <a href="https://carleton.ca/" rel="noopener noreferrer" target="_blank">Carleton University</a>, in Ottawa, he found his way to physics because he was interested in math but preferred the practicality of science.</p><p>While pursuing a master’s degree in physics, also at Carleton, he worked on a water-purification system for the Sudbury Neutrino Observatory, an underground observatory then being built at the bottom of a nickel mine. He would wake up at 4:30 in the morning to drive up to the site, descend 2 kilometers, then enter one of the world’s <a href="https://www.snolab.ca/about/about-snolab/" rel="noopener noreferrer" target="_blank">deepest clean-room facilities</a> to work on the project. The system managed to achieve one atom of impurity per cubic meter of water, “which is pretty insane,” DeKok says.</p><p>But after his master’s degree, DeKok decided to take a different route. Although he found nuclear physics interesting, he says he didn’t see it as his life’s work. Meanwhile, the Ph.D. students he knew were “fanatical about physics.” He had kept up his computing skills through his education, which involved plenty of programming, and decided to look for jobs at computing companies. “I was out of physics. That was it.”</p><p>Still, physics taught him valuable lessons. For one, “You have to understand the big picture,” DeKok says. “The ability to tell the big-picture story in standards, for example, is extremely important.” This skill helps DeKok explain to standards bodies how a protocol acts as one link in the entire chain of events that needs to occur when a user wants to access the internet.</p><p>He also learned that “methods are more important than knowledge.” It’s easy to look up information, but physics taught DeKok how to break down a problem into manageable pieces to come up with a solution. “When I was eventually working in the industry, the techniques that came naturally to me, coming out of physics, didn’t seem to be taught as well to the people I knew in engineering,” he says. “I could catch up very quickly.”</p><h2>Founding FreeRADIUS</h2><p>In 1996, DeKok was hired as a software developer at a company called Gandalf, which made equipment for ISDN, a precursor to broadband that enabled digital transmission of data over telephone lines. Gandalf went under about a year later, and he joined CryptoCard, a company providing hardware devices for two-factor authentication.</p><p>While at CryptoCard, DeKok began spending more time working with a RADIUS server. When users want to connect to a network, RADIUS acts as a gatekeeper and verifies their identity and password, determines what they can access, and tracks sessions. DeKok moved on to a new company in 1999, but he didn’t want to lose the networking skills he had developed. No other open-source RADIUS servers were being actively developed at the time, and he saw a gap in the market.</p><p>The same year, he started FreeRADIUS in his free time and it “gradually took over my life,” DeKok says. He continued to work on the open-source software as a hobby for several years while bouncing around companies in California and France. “Almost by accident, I became one of the more senior people in the space. Then I doubled down on that and started the business.” He founded NetworkRADIUS (now called InkBridge Networks) in 2008.</p><p>By that point, FreeRADIUS was already being used by <a href="https://www.freeradius.org/about/#usage_statistics" rel="noopener noreferrer" target="_blank">100 million people daily</a>. The company now employs experts in Canada, France, and the United Kingdom who work together to support FreeRADIUS. “I’d say at least half of the people in the world get on the internet by being authenticated through my software,” DeKok estimates. He attributes that growth largely to the software being open source. Initially a way to enter the market with little funding, going open source has allowed FreeRADIUS to compete with bigger companies as an industry-leading product.</p><p>Although the software is critical for maintaining secure networks, most people aren’t aware of it because it works behind the scenes. DeKok is often met with surprise that it’s still in use. He compares RADIUS to a building foundation: “You need it, but you never think about it until there’s a crack in it.”</p><h2>27 Years of Fixes</h2><p>Over the years, DeKok has maintained FreeRADIUS by continually making small fixes. Like using a ratcheting tool to make a change inch by inch, “you shouldn’t underestimate that ratchet effect of tiny little fixes that add up over time,” he says.</p><p>He’s seen the project through minor patches and more significant fixes, like when researchers <a href="https://www.freeradius.org/vul_notifications/2024/07/09/blastradius.html" rel="noopener noreferrer" target="_blank">exposed a widespread vulnerability</a> DeKok had been trying to fix since 1998. He also watched a would-be successor to the network protocol, <a href="https://www.sciencedirect.com/topics/engineering/diameter-protocol" rel="noopener noreferrer" target="_blank">Diameter</a>, rise and fall in popularity in the 2000s and 2010s. (Diameter gained traction in mobile applications but has gradually been phased out in the <a href="https://spectrum.ieee.org/telecom-experts-plot-a-path-to-5g" target="_self">shift to 5G</a>.) Though Diameter offers improvements, RADIUS is far simpler and already widely implemented, giving it an edge, DeKok explains.</p><p>And he remains confident about its future. “People ask me, ‘What’s next for RADIUS?’ I don’t see it dying.” Estimating that billions of dollars of equipment run RADIUS, he says, “It’s never going to go away.”</p><p>About his own career, DeKok says he plans to keep working on FreeRADIUS, exploring new markets and products. “I never expected to have a company and a lot of people working for me, my name on all kinds of standards, and customers all over the world. But it worked out that way.”</p><p><em>This article appears in the March 2026 print issue as “<span>Alan DeKok</span>.”</em></p>]]></description><pubDate>Mon, 16 Feb 2026 14:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/network-security-engineer-alan-dekok</guid><category>Typedepartments</category><category>Network-security</category><category>Physics</category><category>Network-servers</category><category>Cybersecurity</category><category>Open-source</category><dc:creator>Gwendolyn Rak</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/person-in-blue-suit-examines-illuminated-server-racks-in-a-dimly-lit-room.png?id=64440487&amp;width=980"></media:content></item><item><title>Startups Are Betting on Orbital Growth for Advanced Electronics</title><link>https://spectrum.ieee.org/in-orbit-manufacturing</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-group-of-people-in-blue-bunny-suits-pose-around-a-satellite-manufacturing-model-in-their-lab.jpg?id=63305822&width=2000&height=1500&coordinates=166%2C0%2C167%2C0"/><br/><br/><p>This past December, the U.K. startup <a href="https://www.spaceforge.com/" rel="noopener noreferrer" target="_blank">Space Forge</a> turned on an orbital furnace aboard its ForgeStar-1 satellite, producing a stream of superhot plasma that could someday enable production of near-ideal semiconductor crystals in orbit. Hailed as a breakthrough in orbital manufacturing, the milestone is a first for a free-flying commercial satellite and a payload not operated by humans.</p><p>Founded in 2018, Space Forge is one of several companies founded on the premise that made-in-space materials can help bring about ultra-efficient next-generation <a href="https://spectrum.ieee.org/tag/electronics" target="_self">electronics</a>, ultrafast <a href="https://spectrum.ieee.org/tag/optical-networks" target="_self">optical networks</a>, and breakthroughs in pharmaceutical research. Space Forge’s flying furnace is specifically designed to make seed crystals that would be used later, on Earth, to produce substrates of gallium and aluminum nitride or silicon carbide for high-performance power devices.</p><p>Semiconductors have been made in space before. <a href="https://osf.io/preprints/osf/d6ar4_v1" rel="noopener noreferrer" target="_blank">Astronauts grew crystals of indium antimonide and germanium </a>in the 1970s aboard the <a href="https://www.nasa.gov/skylab/" rel="noopener noreferrer" target="_blank">Skylab space station</a>, and similar experiments have been taking place on the International Space Station. Between 1973 and 2016, around 160 semiconductor crystals were grown in microgravity aboard various spacecraft, according to <a href="https://www.nature.com/articles/s41526-024-00410-7#:~:text=Semiconductor%20crystals%20grown%20in%20microgravity,%2C3%2C4%2C5.">a meta-analysis </a>study published in the journal <em>Nature</em> in 2024. The study found that 86 percent of those space-grown crystals grew larger, were more uniform, and showed better performance than their counterparts grown on Earth.</p><p>“There is potential for significant energy savings, perhaps as much as 50 percent within large infrastructure installations such as 5G towers,” says <a href="https://youtu.be/JipK7fqCIyQ?si=dL4j12Pvxng6PgLg" rel="noopener noreferrer" target="_blank">Joshua Western</a>, Space Forge’s cofounder and CEO, in an interview.</p><p><a href="https://www.tamus.edu/research/wp-content/uploads/sites/23/2024/08/Putna-Bio.pdf" rel="noopener noreferrer" target="_blank">E. Steve Putna</a>, the director of the <a href="https://chips.tamus.edu/" rel="noopener noreferrer" target="_blank">Texas A&M Semiconductor Institute</a>, says that “space-grown crystals have demonstrated significantly higher electron mobility,” which could translate to a 20 to 40 percent increase in switching efficiency compared with Earth-grown counterparts. (Putna has no involvement with Space Forge.)</p><h2>Why Space Fabrication Makes Better Crystals </h2><p>On Earth, most semiconductors today are <a href="https://www.wevolver.com/article/how-are-semiconductors-made-a-comprehensive-guide-to-semiconductor-manufacturing" rel="noopener noreferrer" target="_blank">fabricated</a> using highly pure materials such as silicon, <a href="https://spectrum.ieee.org/tag/gallium-arsenide" target="_self">gallium arsenide</a>, or <a href="https://spectrum.ieee.org/silicon-carbide" target="_self">gallium nitride</a>, produced by depositing vaporized precursor chemicals layer by layer on <a href="https://inquivixtech.com/what-is-semiconductor-substrates/" rel="noopener noreferrer" target="_blank">substrates</a> inside a vessel called a reactor. The fabrication processes are very tightly controlled, but the resulting semiconductor crystals nevertheless have small amounts of impurities and defects, which can cause heat during operation and slightly reduce performance.</p><p>Western says that the superior vacuum in space does away with those impurities, thus increasing the quality of the crystals produced. ”For example, if you’re worried about nitrogen interfering with your growth process, on Earth [in a vacuum chamber] nitrogen might be present at concentration of around 10 to the –11” (10 to the power of minus 11, or 1 x 10<sup>-9</sup> percent). “In space, above 500 kilometers altitude, it’s naturally present at 10 to the –22.”</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="Satellite transporter floating in space around Earth." class="rm-shortcode" data-rm-shortcode-id="f3e1c6350fb4cb4cec60e64b791986b4" data-rm-shortcode-name="rebelmouse-image" id="498be" loading="lazy" src="https://spectrum.ieee.org/media-library/satellite-transporter-floating-in-space-around-earth.jpg?id=63305931&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">ForgeStar-1 was photographed during its successful test.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Space Forge</small></p><p>Moreover, the <a href="https://www.nasa.gov/centers-and-facilities/glenn/what-is-microgravity/" target="_blank">microgravity</a> environment gives the crystal growth process a “better head start,”<strong> </strong>he says, creating uniform conditions in which the crystals form. “On Earth, you have trouble that, perhaps, some crystals grow around the interior of the reactor and not in other parts because the process between hot and cold is influenced by gravity,” Western says. “Microgravity effectively prevents convection from taking place, so you get a continually uniform deposition area.”</p><p>Such advantages are responsible for the enhanced performance and efficiency of the resulting electronic components made with space-grown crystals, says Western. “The thermal performance of a semiconductor is directly driven by how good its lattice structure is and how good its purity is,” he says. “If you can change those things, then you are able to produce semiconductors that require much less cooling because they run at a lower temperature, which allows you to significantly reduce energy consumption. Or you can trade off that significant reduction in energy consumption for significantly higher output.”</p><p>Putna says the decreased internal resistance and reduced heat generation that results from fewer defects in the semiconductor structure could be “a game changer for <a href="https://spectrum.ieee.org/tag/data-centers" target="_self">AI data centers</a> where cooling costs are a primary bottleneck” for wider deployment. In power electronics, more perfect crystals can allow a smaller chip to handle higher voltages without failing, he adds.<a href="https://www.businesswire.com/news/home/20260114123643/en/Voyager-Secures-Breakthrough-Patent-for-Orbital-Optical-Communications-Manufacturing" target="_blank"></a></p><h2>Will the Performance Boost Justify the High Costs? </h2><p>Still, launching stuff into space and returning it back to Earth is expensive. Currently, <a href="https://www.spacex.com/" target="_blank">SpaceX</a>’s Falcon 9 launches payloads to low Earth orbit for an estimated US <a href="https://www.mckinsey.com/industries/aerospace-and-defense/our-insights/space-launch-are-we-heading-for-oversupply-or-a-shortfall" target="_blank">$1,500 per kilogram.</a> Materials made on the ISS can be returned to Earth on SpaceX’s <a href="https://www.spacex.com/vehicles/dragon" target="_blank">Cargo Dragon</a> capsule, but demand for its services is high and availability limited.</p><p>Western says the seed crystals Space Forge will create in space will be further sprouted in terrestrial foundries while passing on their out-of-this-world qualities. From a single kilogram of space-grown semiconductor, manufacturers on Earth will grow tonnes of high-performance material, he says. “There will be a level of degradation over time and over generations of growth,” he concedes, “but it will be multiple growth runs before the quality degrades to the point of the current state of the art.”</p><p>Putna insists that “if a space-grown substrate increases the yield of a $10,000 high-end AI processor from 50 percent to 90 percent or allows a quantum computer to function closer to room temperature rather than near absolute zero, the launch cost becomes a negligible fraction of the total value created.”</p><p>Not everyone is convinced. <a href="https://iwrc.ieeeusa.org/speakers/matt-francis/" target="_blank">Matt Francis</a>, CEO of the electronics company <a href="https://www.ozarkic.com/" target="_blank">Ozark Integrated Circuits</a>, says that the price of silicon substrates plummeted over the past few years, making infrastructure operators less likely to buy costly space-grown crystals. And terrestrial fabrication technologies, he points out, keep improving. Ozark specializes in rugged ICs for aerospace and other applications.</p><p>“While I remember paying $20K a wafer in the early days, we are down in the hundreds-of-dollars range in volume markets like power,” Francis says. “When they were a prized commodity, maybe sending to space made sense. While the cost of space is decreasing, it’s not decreasing faster than the cost of producing wafers.” <span>In some cases, Francis notes, it might turn out to be possible to get performance superior to a space-grown device by using multiple conventional devices, and at lower overall cost.</span></p><p>Space Forge has yet to test its return technology. Although the ForgeStar-1 satellite will deploy a novel heat shield during its de-orbit maneuver later this year, this spacecraft is designed only to put the orbital furnace through its paces and prove it can “repeatedly create and maintain the manufacturing environment required for the chemistry process” needed to grow superconductor crystals, according to Western.</p><p>The satellite will ultimately perish during its return to Earth, meaning Space Forge will only get its first batch of space-grown crystals back home with the follow-up mission, expected to launch at some point next year. The amount of material brought back will be at best a few kilograms.</p><p>Other companies are betting on the potential of space-grown crystals. Colorado-based Voyager Technologies <a href="https://www.businesswire.com/news/home/20260114123643/en/Voyager-Secures-Breakthrough-Patent-for-Orbital-Optical-Communications-Manufacturing" target="_blank">recently patented a new method</a> for growing crystals of novel fiber-optic materials in orbit, which the company claims could significantly speed up data transmission. London-headquartered <a href="https://www.acmespacehub.com/hyperion/zima" target="_blank">ACME Space</a> wants to test its balloon-launched orbital factory Hyperion later this year, also eyeing the semiconductor, pharmaceuticals, and optical-fiber markets. California-based <a href="https://www.prnewswire.com/news-releases/varda-announces-187-million-in-series-c-funding-to-make-medicines-in-space-302502096.html" target="_blank">Varda Industries raised</a> an impressive $329 million last year to begin manufacturing pharmaceuticals in space. Since 2023, the company has conducted multiple orbital flights to test its return technology and is planning additional tests this year. <a href="https://www.globalinsightservices.com/reports/in-space-manufacturing-market/" target="_blank">Some analysts</a> estimate the in-orbit manufacturing market could reach $28.19 billion by 2034. <del> </del></p><p>Amid this enthusiasm, <a href="https://thebutlercollegian.com/2021/02/a-word-with-butler-chemistry-professor-anne-wilson/" target="_blank">Anne Wilson</a>, a professor of biochemistry at Butler University, in Indianapolis, and an author of the 2024 <em>Nature</em> metanalysis, sounds a note of caution. “I don’t think that microgravity is going to be ideal for the manufacture of bulk materials,” she says. “However, niche materials for specific applications might be worth the investment.”</p>]]></description><pubDate>Tue, 10 Feb 2026 18:27:48 +0000</pubDate><guid>https://spectrum.ieee.org/in-orbit-manufacturing</guid><category>Electronic-components</category><category>Gallium-nitride</category><category>Semiconductor-production</category><category>Semiconductor-manufacturing</category><category>Silicon-carbide</category><dc:creator>Tereza Pultarova</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-group-of-people-in-blue-bunny-suits-pose-around-a-satellite-manufacturing-model-in-their-lab.jpg?id=63305822&amp;width=980"></media:content></item><item><title>How and When the Memory Chip Shortage Will End</title><link>https://spectrum.ieee.org/dram-shortage</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/sk-hynix-inc-s-12-layer-hbm4-memory-chips-on-display.jpg?id=63918122&width=2000&height=1500&coordinates=0%2C0%2C0%2C0"/><br/><br/><p><span><strong>If it feels these</strong> days as if everything in technology is about AI, that’s because it is. And nowhere is that more true than in the market for computer memory. Demand, and profitability, for the type of DRAM used to feed GPUs and other accelerators in AI data centers is so huge that it’s diverting the supply of memory away from other uses and causing prices to skyrocket. According to <a href="https://counterpointresearch.com/en/insights/Memory-Prices-Surge-Up-to-90-From-Q4-2025" target="_blank">Counterpoint Research</a>, DRAM prices have risen 80 to 90 precent so far this quarter.</span></p><p>The largest AI hardware companies are likely first in line for new memory chips, but that leaves everybody else—makers of PCs, consumer gizmos, and everything else that needs to temporarily store a billion bits—scrambling to deal with scarce supply and inflated prices.</p><p>How did the electronics industry get into this mess, and even more important, how will it get out? <em><em>IEEE Spectrum</em></em> asked economists and memory experts to explain. They say today’s situation is the result of a collision between the DRAM industry’s historic boom-and-bust cycle and an AI hardware infrastructure build-out that’s without precedent in its scale. And, barring some major collapse in the AI sector, it will take years for new capacity and new technology to bring supply in line with demand. Prices might stay high even then.</p><p>To understand both ends of the tale, you need to know the main culprit in the supply-and-demand swing: high-bandwidth memory, or HBM.</p><h2>What is HBM?</h2><p>HBM is the DRAM industry’s attempt to short-circuit the slowing pace of Moore’s Law by using 3D chip-packaging technology. Each HBM chip is made up of as many as 12 thinned-down DRAM chips called dies. Each die contains a number of vertical connections called through-silicon vias. The dies are piled atop each other and connected by arrays of microscopic solder balls aligned to the TSVs. This DRAM tower—well, at about 750 micrometers thick, it’s more of a brutalist office block than a tower—is then stacked atop what’s called the base die, which shuttles bits between the memory dies and the processor.</p><p>This complex piece of technology is then set within a millimeter of a GPU or another AI accelerator, to which it is linked by as many as 2,048 µm-scale connections. HBMs are attached on two sides of the processor, and the GPU and memory are packaged together as a single unit.</p><p>The idea behind such a tight, highly connected squeeze with the GPU is to knock down what’s called <a href="https://spectrum.ieee.org/ai-and-memory-wall" target="_self">the memory wall</a>. That’s the barrier in energy and time of bringing the terabytes per second of data needed to run large language models into the GPU. <a href="https://spectrum.ieee.org/ai-models-locally" target="_self">Memory bandwidth</a> is a key limiter to how fast LLMs can run.</p><p>As a technology, HBM has been around for <a href="https://spectrum.ieee.org/chipmakers-push-memory-into-the-third-dimension" target="_self">more than 10 years</a>, and DRAM makers have been busy boosting its capability.</p><div class="flourish-embed flourish-chart" data-src="visualisation/27466742?1509099"><script src="https://public.flourish.studio/resources/embed.js"></script><noscript><img alt="chart visualization" src="https://public.flourish.studio/visualisation/27466742/thumbnail" width="100%"/></noscript></div><p>As the size of AI models has grown, so has HBM’s importance to the GPU. But that’s come at a cost. <a href="https://newsletter.semianalysis.com/p/scaling-the-memory-wall-the-rise-and-roadmap-of-hbm" target="_blank">SemiAnalysis estimates</a><strong> </strong>that HBM generally costs three times as much as other types of memory and constitutes 50 percent or more of the cost of the packaged GPU.</p><h2>Origins of the memory chip shortage</h2><p>Memory- and storage-industry watchers agree that DRAM is a highly cyclical industry with huge booms and devastating busts. With new fabs costing US $15 billion or more, firms are extremely reluctant to expand and may only have the cash to do so during boom times, explains <a href="https://www.linkedin.com/in/thomas-coughlin-41a65/" target="_blank">Thomas Coughlin</a>, a storage and memory expert and president of <a href="https://tomcoughlin.com/" target="_blank">Coughlin Associates</a>. But building such a fab and getting it up and running can take 18 months or more, practically ensuring that new capacity arrives well past the initial surge in demand, flooding the market and depressing prices.</p><p>The origins of today’s cycle, says Coughlin, go all the way back to the <a href="https://spectrum.ieee.org/chip-shortage" target="_self">chip supply panic surrounding the COVID-19 pandemic</a>. To avoid supply-chain stumbles and support the rapid shift to remote work, hyperscalers—data-center giants like Amazon, Google, and Microsoft—bought up huge inventories of memory and storage, thus boosting prices, he notes.</p><p>But then supply became more regular and data-center expansion fell off in 2022, causing memory and storage prices to plummet. This recession continued into 2023, and even resulted in big memory and storage companies such as Samsung cutting production by 50 percent to try to keep prices from going below the costs of manufacturing, says Coughlin. It was a rare and fairly desperate move, because companies typically have to run plants at full capacity just to earn back their value.<span></span></p><p>After a recovery began in late 2023, “all the memory and storage companies were very wary of increasing their production capacity again,” says Coughlin. “Thus there was little or no investment in new production capacity in 2024 and through most of 2025.”</p><div class="flourish-embed flourish-chart" data-src="visualisation/27468004?1509099"><script src="https://public.flourish.studio/resources/embed.js"></script><noscript><img alt="chart visualization" src="https://public.flourish.studio/visualisation/27468004/thumbnail" width="100%"/></noscript></div><h2>The AI data-center boom</h2><p>That lack of new investment is colliding headlong with a huge boost in demand from new data centers. Globally, there are <a href="https://spectrum.ieee.org/data-center-growth" target="_self">nearly 2,000 new data centers</a> either planned or under construction right now, according to Data Center Map. If they’re all built, it would represent a 20 percent jump in the global supply, which stands at around 9,000 facilities now.</p><p>If the current build-out continues at pace, McKinsey predicts companies will spend <a href="https://programs.com/resources/data-center-statistics/" target="_blank">$7 trillion by 2030</a>, with the bulk of that—$5.2 trillion—going to AI-focused data centers. Of that chunk, $3.3 trillion will go toward servers, data storage, and network equipment, the firm predicts.</p><p>The biggest beneficiary so far of the AI data-center boom is unquestionably the GPU maker Nvidia. Revenue for its data-center business went from <a href="https://ycharts.com/indicators/nvidia_corp_nvda_data_center_revenue_quarterly" target="_blank">barely a billion in the final quarter of 2019 to $51 billion in the quarter that ended in October 2025</a>. Over this period, its server GPUs have demanded not just more and more gigabytes of DRAM but an increasing number of DRAM chips. The recently released B300 uses eight HBM chips, each of which is a stack of 12 DRAM dies. Competitors’ use of HBM has largely mirrored Nvidia’s. AMD’s MI350 GPU, for example, also uses eight 12-die chips.</p><div class="flourish-embed flourish-chart" data-src="visualisation/27459660?1509099"><script src="https://public.flourish.studio/resources/embed.js"></script><noscript><img alt="chart visualization" src="https://public.flourish.studio/visualisation/27459660/thumbnail" width="100%"/></noscript></div><p>With so much demand, an increasing fraction of the revenue for DRAM makers comes from HBM. Micron—the No. 3 producer behind SK Hynix and Samsung—reported that <a href="https://investors.micron.com/static-files/8791eb80-8263-4c6f-aa74-fdd03fbbb027" target="_blank">HBM and other cloud-related memory</a> went from being 17 percent of its DRAM revenue in 2023 to nearly 50 percent in 2025.</p><p>Micron predicts the total market for HBM will grow from $35 billion in 2025 to $100 billion by 2028—a figure larger than the entire DRAM market in 2024, CEO <a href="https://www.linkedin.com/in/sanjay-mehrotra/" target="_blank">Sanjay Mehrotra</a> <a href="https://investors.micron.com/static-files/088991c5-a249-4f66-a0a6-258d9b66f3f9" target="_blank">told analysts in December</a>. It’s reaching that figure two years earlier than Micron had previously expected. Across the industry, demand will outstrip supply “substantially…for the foreseeable future,” he said.</p><div class="flourish-embed flourish-chart" data-src="visualisation/27481588?1509099"><script src="https://public.flourish.studio/resources/embed.js"></script><noscript><img alt="chart visualization" src="https://public.flourish.studio/visualisation/27481588/thumbnail" width="100%"/></noscript></div><h2>Future DRAM supply and technology</h2><p>“There are two ways to address supply issues with DRAM: with innovation or with building more fabs,” explains <a href="https://www.linkedin.com/in/mina-kim-37449b/" target="_blank">Mina Kim</a>, an economist with Mkecon Insights. “As <a href="https://spectrum.ieee.org/micron-dram" target="_self">DRAM scaling</a> has become more difficult, the industry has turned to advanced packaging… which is just using more DRAM.”</p><p>Micron, Samsung, and SK Hynix combined make up the vast majority of the memory and storage markets, and all three have new fabs and facilities in the works. However, these are unlikely to contribute meaningfully to bringing down prices.</p><p><strong>Micron</strong> is in the process of <a href="https://investors.micron.com/news-releases/news-release-details/micron-breaks-ground-advanced-wafer-fabrication-facility" target="_blank">building an HBM fab</a> in Singapore that should be in production in 2027. And it is <a href="https://investors.micron.com/news-releases/news-release-details/micron-signs-letter-intent-purchase-tongluo-site-begin-0" target="_blank">retooling a fab</a> it purchased from PSMC in Taiwan that will begin production in the second half of 2027. Last month, Micron <a href="https://investors.micron.com/news-releases/news-release-details/micron-celebrates-official-groundbreaking-new-york-megafab-site" target="_blank">broke ground</a> on what will be a DRAM fab complex in Onondaga County, N.Y. It will not be in full production until 2030.</p><p><strong>Samsung</strong> plans to <a href="https://www.chosun.com/english/industry-en/2025/11/16/U5VKNZCSYBCONMRCXDZ45MCIXQ/" target="_blank">start producing</a> at a new plant in Pyeongtaek, South Korea, in 2028.</p><p><strong>SK Hynix</strong> is building <a href="https://www.skhynix.com/westlafayette.IN/" target="_blank">HBM and packaging</a> facilities in West Lafayette, Ind., set to begin production by the end of 2028, and an HBM fab it’s <a href="https://www.cnbc.com/2026/01/13/sk-hynix-invest-13-billion-new-fab-memory-chip-shortage-advanced-packaging-ai-memory.html" target="_blank">building in Cheongju</a>, South Korea, should be complete in 2027.</p><p>Speaking of his sense of the DRAM market, <a href="https://newsroom.intel.com/biography/lip-bu-tan" target="_blank">Intel CEO Lip-Bu Tan</a><strong> </strong>told attendees at the <a href="https://newsroom.cisco.com/c/r/newsroom/en/us/a/y2026/m02/ai-summit.html" target="_blank">Cisco AI Summit</a> last week, “There’s no relief until 2028.”</p><p>With these expansions unable to contribute for several years, other factors will be needed to increase supply. “Relief will come from a combination of incremental capacity expansions by existing DRAM leaders, yield improvements in advanced packaging, and a broader diversification of supply chains,” says <a href="https://www.electronics.org/meet-shawn-dubravac-ipcs-chief-economist" rel="noopener noreferrer" target="_blank">Shawn DuBravac</a>, chief economist for the <a href="https://www.electronics.org/" rel="noopener noreferrer" target="_blank">Global Electronics Association</a> (formerly the IPC). “New fabs will help at the margin, but the faster gains will come from process learning, better [DRAM] stacking efficiency, and tighter coordination between memory suppliers and AI chip designers.”</p><p>So, will prices come down once some of these new plants come on line? Don’t bet on it. “In general, economists find that prices come down much more slowly and reluctantly than they go up. DRAM today is unlikely to be an exception to this general observation, especially given the insatiable demand for compute,” says Kim.</p><p>In the meantime, technologies are in the works that could make HBM an even bigger consumer of silicon. The standard for HBM4 can accommodate 16 stacked DRAM dies, even though today’s chips only use 12 dies. Getting to 16 has a lot to do with the chip-stacking technology. Conducting heat through the HBM “layer cake” of silicon, solder, and support material is a key limiter to going higher and in <a href="https://spectrum.ieee.org/hbm-on-gpu-imec-iedm" target="_self">repositioning HBM inside the package</a> to get even more bandwidth.</p><p>SK Hynix claims a heat-conduction advantage through a manufacturing process called advanced <a href="https://news.skhynix.com/rulebreaker-revolutions-mr-muf-unlocks-hbm-heat-control/" rel="noopener noreferrer" target="_blank">MR-MUF (mass reflow molded underfill)</a>. Further out, an alternative chip-stacking technology called <a href="https://spectrum.ieee.org/hybrid-bonding" target="_self">hybrid bonding</a> could help heat conduction by reducing the die-to-die vertical distance essentially to zero. In 2024, researchers at Samsung proved they could produce a 16-high stack with hybrid bonding, and they suggested that <a href="https://spectrum.ieee.org/hybrid-bonding" target="_self">20 dies was not out of reach</a>.</p><p><em>This post was corrected on 9 March. Previously it misrepresented the position of large AI hardware makers, and it misstated the scale of datacenter hardware market.</em></p><p><em>This article appears in the April 2026 print issue as the feature “<span>AI </span>is a Memory </em><span><em>Hog.”</em></span></p><p><span><em>The sidebar from that article is here “<a href="https://spectrum.ieee.org/ram-shortage-price-increase" target="_blank">Low-Cost Computers Nearly Double in Price as RAM Shortage Hits.</a>“</em></span></p>]]></description><pubDate>Tue, 10 Feb 2026 14:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/dram-shortage</guid><category>Ai-hardware</category><category>Computer-memory</category><category>Dram</category><category>Hbm</category><category>Data-centers</category><dc:creator>Samuel K. Moore</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/sk-hynix-inc-s-12-layer-hbm4-memory-chips-on-display.jpg?id=63918122&amp;width=980"></media:content></item></channel></rss>