<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:media="http://search.yahoo.com/mrss/"><channel><title>IEEE Spectrum</title><link>https://spectrum.ieee.org/</link><description>IEEE Spectrum</description><atom:link href="https://spectrum.ieee.org/feeds/topic/computing.rss" rel="self"></atom:link><language>en-us</language><lastBuildDate>Thu, 30 Apr 2026 13:00:01 -0000</lastBuildDate><image><url>https://spectrum.ieee.org/media-library/eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpbWFnZSI6Imh0dHBzOi8vYXNzZXRzLnJibC5tcy8yNjg4NDUyMC9vcmlnaW4ucG5nIiwiZXhwaXJlc19hdCI6MTgyNjE0MzQzOX0.N7fHdky-KEYicEarB5Y-YGrry7baoW61oxUszI23GV4/image.png?width=210</url><link>https://spectrum.ieee.org/</link><title>IEEE Spectrum</title></image><item><title>The Fog, a New Encrypted Cloud Platform, Rolls In</title><link>https://spectrum.ieee.org/the-fog-cloud-encryption</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/illustration-of-a-workflow-application-floating-on-a-cloud-while-tethered-to-an-encrypted-chip-below.jpg?id=65521062&width=1245&height=700&coordinates=0%2C469%2C0%2C469"/><br/><br/><p>Most cloud computing services encrypt data in transit and at rest. But that data still needs to be decrypted before cloud servers or virtual machines can perform any kind of computation on it. This risks exposing data—especially sensitive information such as financial transactions or medical records—during processing. This is where <a href="https://niobium.co/platform" rel="noopener noreferrer" target="_blank">The Fog</a> comes in.</p><p>Launched in early April by chip startup <a href="https://niobium.co/" rel="noopener noreferrer" target="_blank">Niobium</a>, The Fog is an encrypted cloud platform. It follows a client-server architecture, where a person or organization (the client) can encrypt data or workloads locally using their own private keys and deploy the encrypted data or workloads to The Fog (the server) without sharing their keys. These private keys remain with data owners, and only they can decrypt any results from the platform.</p><p>Much like actual fog obscures everything it envelops, so does the encrypted cloud platform named after it. Yet unlike physical fog that eventually lifts, The Fog keeps data opaque at all times—even as computation happens.</p><p>“The data in our cloud will never be exposed—it’s always encrypted,” says <a href="https://www.linkedin.com/in/barrus" rel="noopener noreferrer" target="_blank">John Barrus</a>, vice president of product at Niobium. “It’s a new category of cloud.”</p><h2>Fully Homomorphic Encryption keeps The Fog secure</h2><p>Beneath The Fog lies a cryptographic technique known as <a href="https://spectrum.ieee.org/tag/homomorphic-encryption" target="_self">fully homomorphic encryption</a>, or FHE, which allows for computing on encrypted data without the need to decrypt it. But FHE is often slow and requires a lot of computing power and memory. </p><p>Niobium aims to address these bottlenecks using mistic, its <a href="https://spectrum.ieee.org/tag/fpga" target="_self">FPGA</a> (field-programmable gate array) chip, which can be reconfigured for FHE after manufacturing. For some applications the company is testing, its <a href="https://spectrum.ieee.org/homomorphic-encryption" target="_self">accelerator hardware</a> runs FHE about twice as fast as today’s GPUs, Barrus says.</p><p>To demonstrate the usability of its encrypted cloud platform, Niobium has developed a handful of template applications “that solve typical problems where you might want to hide the data or keep it encrypted, so people can start there and just try it out,” says Barrus. One such template application involves encrypted semantic search, which queries databases or datasets and returns relevant results based on the context or meaning of the search terms rather than keywords that match them. Both the query and the data source are encrypted, helping ensure data privacy.</p><p>“Let’s say you’re a legal firm, and you have sensitive case documents. You encrypt all those documents and store them encrypted in the cloud,” Barrus says. In this scenario, you can ask questions about the documents using encrypted semantic search “and get pointers to those documents back, and then just download and decrypt the documents you need.”</p><h2>Niobium takes FHE from theory to practice</h2><p><a href="https://www.linkedin.com/in/kurt-rohloff/" rel="noopener noreferrer" target="_blank">Kurt Rohloff</a>, cofounder and chief technology officer (CTO) at <a href="https://dualitytech.com/" rel="noopener noreferrer" target="_blank">Duality Technologies</a>, is excited about the prospect of running his company’s privacy-enhancing software products on The Fog. Duality provides software that uses FHE, including an <a href="https://spectrum.ieee.org/homomorphic-encryption-llm" target="_self">LLM inference framework</a>. Without a platform like The Fog, users may need to purchase dedicated FHE acceleration hardware, he says. But “the Niobium encrypted cloud platform allows users to rapidly scale their use of FHE-protected computing [and] get much more value from their data,” he says.</p><p>Echoing the sentiment is <a href="https://www.linkedin.com/in/rashmi-agrawal-9a0601133" rel="noopener noreferrer" target="_blank">Rashmi Agrawal</a>, cofounder and CTO at <a href="https://www.ciphersoniclabs.io/" rel="noopener noreferrer" target="_blank">CipherSonic Labs</a>, a company building FHE-powered encrypted AI infrastructure. “Platforms like Niobium are important because they help move FHE from theory into deployable infrastructure,” she says. “An encrypted cloud platform built on FHE fundamentally changes the trust model of cloud computing. This significantly reduces exposure to data leakage, insider threats, and compliance risks while enabling organizations to safely process highly sensitive data in the cloud.”</p><p>However, Agrawal points out that despite FHE’s rapid progress, there are still practical challenges. These include performance overheads for complex tasks or workloads that need to be completed with low latency, as well as filling in skills gaps for software developers who have no FHE knowledge or experience. “Building FHE-compatible applications often requires rethinking traditional approaches. The ecosystem is still maturing as tooling, standards, and interoperability continue to evolve,” she adds.</p><p>Barrus acknowledges these hurdles. “I think the real challenge is large language models with a lot of matrix and vector multiplications. We have to be fast enough that you’re not waiting minutes for every token but seconds or so. That’s going to be much harder to solve,” he says.</p><p>In terms of equipping developers without any FHE background, Niobium hopes to make The Fog more accessible by providing a tech stack composed of a compiler, software development kit, documentation, and other training materials. “If we can bring FHE computation to more people, then more people can develop privacy-preserving applications,” says Barrus.</p><p>The Fog is currently available in private beta, with Niobium targeting May or June for a public launch. The company is also developing an application-specific integrated circuit for its encrypted cloud platform that Barrus says will be up to 25 times as fast as a GPU, depending on the application.</p><p>“What we’re trying to do is create value from encrypted data,” he says. “Our vision is that data never has to be exposed to be useful.”</p>]]></description><pubDate>Thu, 30 Apr 2026 13:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/the-fog-cloud-encryption</guid><category>Homomorphic-encryption</category><category>Data-privacy</category><category>Cloud-computing</category><category>Hardware-acceleration</category><dc:creator>Rina Diane Caballar</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/illustration-of-a-workflow-application-floating-on-a-cloud-while-tethered-to-an-encrypted-chip-below.jpg?id=65521062&amp;width=980"></media:content></item><item><title>Power Buffer Protects Grid From Data Centers’ Wild Load Swings</title><link>https://spectrum.ieee.org/data-center-power-fluctuation</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/staff-working-behind-computers-in-the-control-center.jpg?id=66648750&width=1245&height=700&coordinates=0%2C62%2C0%2C63"/><br/><br/><p>As more AI data centers come on line, concerns are rising about their effects on the grid, and it’s not just the amount of power they consume. They tend to have huge swings in power use, surging up and down by 70 percent or more in milliseconds. Traditional electricity infrastructure isn’t designed to deal with that kind of load fluctuation. </p><p>To address the problem, researchers are developing power electronics systems that sit between the data center and the grid to act as a buffer and even as a grid helper in times of need. One such system, developed by the Miami-based company <a href="https://www.on.energy/" rel="noopener noreferrer" target="_blank">ON.energy</a>, is being implemented across 3 gigawatts’ worth of projects, and has sailed through a battery of tests at the U.S. National Lab of the Rockies (NLR).</p><p>In the tests, ON.energy’s system sat between a simulated data center and a simulated grid. The system successfully protected the data center from grid instability and also safeguarded the grid from the major load swings generated by the data center. The company’s technology involves a bidirectional uninterruptible power supply (UPS) that it calls AI UPS.</p><p>Such grid buffers are becoming increasingly important as AI facilities expand to gigawatt scale and beyond. Utilities have major concerns about both the amount of power demanded by these data centers and their potential to create system instability due to wild variations in loads. Innovations are needed to help data centers become better grid citizens, and shorten the amount of time they must wait to connect to the grid.</p><h2>AI Data Centers and Grid Stability</h2><p>UPS systems have been used for decades to protect data centers from grid events. If frequency varies suddenly or power is lost, these unidirectional systems provide almost instantaneous, short-term backup power to the equipment inside the data center. Because servers can’t tolerate more than minor deviations, UPS electronics also clean up low-quality power, such as voltage spikes or sags and frequency deviation.</p><p>UPS has served data centers well. But the scale of modern facilities packed with graphics processing units (GPUs) changes the game. Instead of data centers whose size is measured in tens of megawatts, <a href="https://spectrum.ieee.org/5gw-data-center" target="_self">AI facilities are reaching up to 5 GW</a>. They still require the type of protection afforded by UPS, but their massive scale and load volatility pose dangers to the grid. </p><p>During a minor grid fault in Virginia in 2025, for example, several data centers tripped offline, causing <a href="https://www.datacenterdynamics.com/en/news/virginia-narrowly-avoided-power-cuts-when-60-data-centers-dropped-off-the-grid-at-once/" rel="noopener noreferrer" target="_blank">1.5 GW to drop off the grid</a> simultaneously. This caused panic for the system operator, who had to act fast to balance the system and avoid a major power outage.</p><p>In addition to major changes in overall load, AI data centers can generate short-lived, high-voltage, or high-current disturbances known as grid transients. They may only last microseconds, but they can break down insulation, overheat transformers, cause electrical arcing, start fires, and destabilize an entire grid.</p><p>“The scale of modern data centers could lead to load swings of 1 GW multiple times per minute, which creates frequency variations and oscillations that the grid can’t handle,” says <a href="https://www.linkedin.com/in/ricardodeazevedo/" rel="noopener noreferrer" target="_blank">Ricardo de Azevedo</a>, CTO at ON.energy. </p><p>These problems have given utilities and government authorities pause. Some authorities in the United States and parts of Europe are implementing moratoriums on new data centers or instituting rules that place responsibility for grid conditions onto the data center.</p><p><a href="https://www.gtlaw.com/en/insights/2026/3/texas-senate-bill-6-update-what-data-centers-large-load-customers-should-know-about-proposed-interconnection-standards" rel="noopener noreferrer" target="_blank">Texas Senate Bill 6</a>, for example, requires new data centers to pay a share of any new grid infrastructure needed by their facilities. Additional requirements for voltage ride-through—the equipment’s ability to continue operating during power disruptions—are currently being formulated in accordance with this bill. Such rules aim to prevent large data centers from tripping offline suddenly or overwhelming the grid due to severe load variability from AI workloads. </p><p>“One of our customers in Texas that is building a 1-GW campus is now being required by the local grid authority to include voltage ride-through,” Azevedo says. </p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Two men monitoring power grid data displayed on a video wall." class="rm-shortcode" data-rm-shortcode-id="3d7f1a685653b94a940be5dca341e269" data-rm-shortcode-name="rebelmouse-image" id="f5d69" loading="lazy" src="https://spectrum.ieee.org/media-library/two-men-monitoring-power-grid-data-displayed-on-a-video-wall.jpg?id=66648905&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">NLR engineers Przemyslaw Koralewicz [left] and Shahil Shah monitor the results of a simulation of ON.energy’s AI UPS in the control center at the NLR Flatirons Campus.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Agata Bogucka/NLR</small></p><h2>Bidirectional UPS </h2><p>ON.energy’s 3.5-megawatt units consist of a power conversion system (PCS), batteries to store energy and act as an energy reservoir or buffer, another PCS, and a transformer. The batteries can provide up to eight hours of backup power, depending on the size of the data center. ON.energy sources this equipment from established manufacturers and adds its own software and controls.</p><p>The latest PCS units are bidirectional, acting as the interface between the grid, the batteries, and the data center. They convert between the alternating current (AC) from the grid, the direct current (DC) stored in the batteries, and the AC delivered to the data-center load, ensuring power quality and optimal flow when feeding an AI facility. In the other direction, the PCS absorbs and smooths transients caused by sudden load swings in the data center that would otherwise disrupt the grid. </p><p>“The batteries act like a reservoir of energy as well as a shock absorber, should there be any disturbances on the grid or from the data center,” says Azevedo.</p><p>ON.energy’s system is housed outside the data center, rather than inside as most UPS systems are, which frees up space internally for more compute resources. Being outside also allows the system to harness more advanced power electronics fed by medium voltage. Traditional UPS, on the other hand, operates on the low voltages needed by data-center computers for safety reasons. </p><p>The company has about 3 GW of these bidirectional AI UPS units either operating or under construction. It expects to commission a system in May for a 1.5-GW AI data center in Texas, according to Azevedo. For such a facility, hundreds of these 3.5-MW units would be required. </p><h2>NLR’s Data Center-Grid Simulator</h2><p>To test its system, ON.energy turned to NLR (formerly known as the National Renewable Energy Laboratory). The facility is likely the <a href="https://www.nlr.gov/news/detail/program/2026/could-a-new-kind-of-power-supply-help-make-data-centers-grid-friendly" target="_blank">only one in the world</a> that can do full-load, bidirectional testing that simulates both grid conditions and variable data-center loads. The facility can test up to 20 MW with voltage levels reaching 13.2 kilovolts. The test consisted of a 7-MW grid simulator that replicates disturbances and voltage ride-through events, and a 20-MW load simulator that reproduces real-world demand dynamics such as those created by an AI data center.</p>Systems like ON.energy’s could become the norm in the coming years. Pilot projects for similar technologies are ongoing in Ireland. Another <a href="https://spectrum.ieee.org/dcflex-data-center-flexibility" target="_self">project in France</a> coordinated by the Electric Power Research Institute (EPRI) is assessing the capabilities of UPS systems through its DC Flex initiative. Results are expected in the coming weeks. Lower voltage versions of this type of bidirectional technology are also under development by <a href="https://www.eaton.com/us/en-us/products/backup-power-ups-surge-it-power-distribution/backup-power-ups/dual-purpose-ups-technology.html" target="_blank">Eaton and Microsoft</a>. <p><br/></p><em><em>This story was updated on 29 April, 2026 to clarify how ON.energy’s power conversion system works.</em></em>]]></description><pubDate>Wed, 29 Apr 2026 14:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/data-center-power-fluctuation</guid><category>Data-center-energy</category><category>Ai-data-centers</category><category>Power-quality</category><category>Power-electronics</category><dc:creator>Drew Robb</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/staff-working-behind-computers-in-the-control-center.jpg?id=66648750&amp;width=980"></media:content></item><item><title>Why the Ideal Magnet Remains Out of Reach</title><link>https://spectrum.ieee.org/rare-earth-free-magnets</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/photo-of-a-technician-working-on-a-cylindrical-machine-containing-brass-colored-components.jpg?id=66525633&width=1245&height=700&coordinates=0%2C469%2C0%2C469"/><br/><br/><p>All over the world, researchers are working on an urgent and surprisingly difficult challenge: creating a cost-effective yet powerful <a href="https://spectrum.ieee.org/best-rare-earth-elements-2025" target="_self">permanent magnet</a> that doesn’t use <a href="https://spectrum.ieee.org/rare-earth-elements-2670490876" target="_self">rare earth elements</a>. Rare earth magnets are essential components of the motors for electric vehicles, heating and cooling systems, robots, tools, and appliances, and they’re also essential for wind turbines, audio speakers, and other systems. A strong magnet that doesn’t use rare earths would be of almost incalculable value, because it would free its users from China’s near-monopoly on rare earth elements and magnets. By circumventing that monopoly, it would almost certainly alter geostrategic calculations and global supply chains in short order.</p><p>Tantalizingly, no physics theories preclude the existence of a powerful and rare-earth-free magnet. And yet, after more than a decade of intensive efforts by many exceptionally bright people, no such magnet has been discovered.</p><p>Now, a small group of researchers in France and the United States has set out to test an intriguing hypothesis—that the problem can be solved with quantum computers. “You need the math of quantum mechanics to solve a problem that lives in the quantum realm,” declares <a href="https://www.linkedin.com/in/theau-peronnin/" rel="noopener noreferrer" target="_blank">Théau Peronnin</a>, CEO of <a href="https://alice-bob.com/" rel="noopener noreferrer" target="_blank">Alice & Bob</a>, a Paris-based quantum computer startup. Alice & Bob is collaborating with <a href="https://www.lanl.gov/" rel="noopener noreferrer" target="_blank">Los Alamos National Laboratory</a> and <a href="https://www.gevernova.com/" rel="noopener noreferrer" target="_blank">GE Vernova</a>, with US $3.9 million in funding from the U.S. Department of Energy’s ARPA-E <a href="https://arpa-e.energy.gov/programs-and-initiatives/view-all-programs/qc3" rel="noopener noreferrer" target="_blank">Quantum Computing for Computational Chemistry</a> program.</p><h2><strong>Why Rare Earth Magnets Still Dominate</strong></h2><p><a href="https://nemad.org/" rel="noopener noreferrer" target="_blank">More than 67,000 compounds</a> are known to have some degree of permanent magnetism. None, however, come close to the reigning permanent-magnet champ, <a href="https://spectrum.ieee.org/the-men-who-made-the-magnet-that-made-the-modern-world" target="_self">neodymium iron boron</a> (NdFeB), which dominates high-power applications.</p><p>For more than 15 years, researchers have used conventional high-performance computers to search for new and powerful magnets. But no commercially successful magnets have come out of that work. Even the best conventional computers aren’t powerful enough to simulate the detailed magnetic properties of a hypothetical permanent magnet.</p><p>To understand why, start with the basics. Permanent magnetism arises in certain crystalline materials when the spins of <a href="https://spectrum.ieee.org/tag/electrons" target="_self">electrons</a> of some of the atoms in the crystal are forced to point in the same direction, either “up” or “down.” The more of these aligned spins, the stronger the magnetism. The ideal atoms are ones that have unpaired electrons swarming around the nucleus in what are known as <a href="https://winter.group.shef.ac.uk/orbitron/atomic_orbitals/3d/index.html" rel="noopener noreferrer" target="_blank">3d orbitals</a>. Tops are iron, with four unpaired 3d electrons, and <a href="https://spectrum.ieee.org/tag/cobalt" target="_self">cobalt</a>, with three.</p><p>But 3d electrons alone are not enough to make superstrong magnets. As researchers discovered decades ago, magnetic strength can be greatly improved by adding to the crystalline lattice atoms with unpaired electrons in the 4f orbital—notably the rare earth elements <a href="https://spectrum.ieee.org/tag/neodymium" target="_self">neodymium</a>, praseodymium, and dysprosium. These 4f electrons enhance a characteristic of the crystalline lattice called magnetic <a href="https://www.stanfordmagnets.com/what-is-magnetic-anisotropy.html" rel="noopener noreferrer" target="_blank">anisotropy</a>—in effect, they promote adherence of the magnetic moments of the atoms to the desired directions in the crystal lattice. That, in turn, can be exploited to achieve high <a href="https://en.wikipedia.org/wiki/Coercivity" rel="noopener noreferrer" target="_blank">coercivity</a>, the essential property that lets a permanent magnet stay magnetized.</p><p class="pull-quote">“The combinatorial space is just ridiculously large. It’s 2 to the—I don’t know—40th or 50th power. It’s absolutely tremendous.”<br/></p><p>The point is that being able to accurately simulate a hypothetical magnet means not only accounting for all those electron orbitals and spin states but also simulating the <em><em>interaction</em></em> of all those electron orbitals and spin states. And that’s really, really hard.</p><p>“Let’s say you have a chain of atoms, each with a single electron in the 1d orbital,” explains Peronnin. “And then you want to understand: If the spin of this one electron is down, how does it affect its neighbors? Would they be more likely to be up or down? And you need to do so for all the electrons in your chain. And then see if the total system has a tendency to align all its electron spins. Or, once you’ve added a bit of thermal noise and an external magnetic field, for example, how much disorder would there be in that chain? And so those are exactly the properties you want to predict.</p><p>“The emergent global properties [such as magnetism] arise from the local behavior of each electron. But each electron’s behavior is highly, highly correlated with how its neighbors behave. And this is what makes the problem extremely difficult, because you cannot treat each of those electrons individually. You need to treat the whole system with all its possible configurations all at once to predict the global properties. And this is where the computing space explodes.</p><p>“You have to consider all the possible superpositions of states of those electrons,” Peronnin continues. “And so here, the combinatorial space is just ridiculously large. It’s 2 to the, I don’t know, 40th or 50th power. It’s absolutely tremendous.”</p><h2>Why Quantum Computers Might Finally Solve This Problem</h2><p>The great potential advantage of quantum computers here is <a href="https://arxiv.org/abs/2405.07222" rel="noopener noreferrer" target="_blank">quantum parallelism</a>, a capability that emerges directly from the qubits that are the heart of a quantum computer. In such a machine, these qubits are entangled with one another. The qubits are also in a state of <a href="https://scienceexchange.caltech.edu/topics/quantum-science-explained/quantum-superposition" rel="noopener noreferrer" target="_blank">superposition</a>, which means that they can embody, in the macro world, certain quantum characteristics of subatomic particles. Namely, they can represent a binary 0 or 1 and also exist in a continuous range of states, each with an associated pair of probabilities—a probability that the bit is 0 and a corresponding probability that it’s 1. And the more there are of these superimposed qubits that are entangled, the more states those qubits can represent: A collection of <em><em>n</em></em> entangled qubits can represent 2<em><em><span><sup>n</sup></span></em></em> states simultaneously. The upshot is that with enough qubits, a quantum computer could handle the stupendous computational challenge of accurately simulating a hypothetical magnetic material.</p><p>How many qubits are enough? Peronnin figures things will start getting interesting when he and his colleagues can build a machine that has 100 logical qubits furnished with a proprietary type of error correction that they have pioneered. He figures that will happen around 2030. (IBM and others have already built quantum computers with <a href="https://spectrum.ieee.org/ibm-condor" target="_self">over 1,000 <em><em>physical</em></em> qubits</a>, but these machines did not have the <a href="https://spectrum.ieee.org/ibm-quantum-error-correction-starling" target="_self">error correction</a> that is the defining characteristic of logical qubits, and none of them ever performed useful work.)</p><p class="pull-quote">A strong magnet that doesn’t use rare earths would be of almost incalculable value.</p><p>Magnetics researchers not involved with the ARPA-E effort are mostly supportive of the project, while noting that progress on quantum computers is notoriously difficult to predict. “This is an interesting approach,” says <a href="https://ceps.unh.edu/person/jiadong-zang" target="_blank">Jiadong Zang</a>, a professor of materials science and director of the materials science program at the University of New Hampshire. “You need some extraordinary approach to find some new structures,” he adds. Zang is part of a group that has been using a large language model to search the magnetics literature for the purpose of creating a database of experimental magnets, called the <a href="https://pubmed.ncbi.nlm.nih.gov/41136402/" target="_blank">Northeast Materials Database for Magnetic Materials</a>.</p><p>“This might be a task that quantum computers could do well,” agrees <a href="https://www.ameslab.gov/directory/matthew-kramer" target="_blank">Matthew Kramer</a>, Distinguished Scientist at <a href="https://www.ameslab.gov/" target="_blank">Ames National Laboratory</a>, in Iowa. (Kramer is working on a project with the U.S. Department of Energy and Fermilab aimed at improving a certain class of qubits.) He cautions, however, that efforts to use conventional computers to identify new magnet materials have often identified new candidates that could not possibly be built in the real world.</p><h2>Microsoft’s Imaginary Magnets Will Probably Stay That Way</h2><p>A recent and highly <a href="https://www.microsoft.com/en-us/research/blog/mattergen-a-new-paradigm-of-materials-design-with-generative-ai/" target="_blank">ambitious project at Microsoft</a>, for example, resulted in a system called <a href="https://www.nature.com/articles/s41586-025-08628-5" target="_blank">MatterGen</a>, which the researchers used to design a range of magnets with “low supply-chain risk.” However, the researchers simplified the problem greatly by focusing on “high magnetic density” alone, without trying to incorporate any of the many other characteristics needed for a magnet to be useful. Taking into account such characteristics, including high coercivity, chemical stability, and cost effectiveness, is a big reason why the challenge quickly becomes computationally intractable. In the end, the researchers did not fabricate any of the magnets identified; it’s not even clear that they could.</p><p>“They had a lot of unusual structures,” Kramer notes. “The real question there is, can any of those actually be synthesized?”</p>At GE Vernova, <a href="https://scholar.google.com/citations?user=E2xhYPAAAAAJ&hl=en" target="_blank">senior scientist Jonathan Owens</a> says a likely best outcome would be for quantum computing to become part of a larger experimental system. “Quantum will be a piece of probably a much larger pipeline where you’re using machine learning or traditional methods to kind of guide what quantum calculations you need to run,” Owens says. “You’ll feed that back into your larger workflow and sort of iterate. But you can explore any space because you’re not restricted to only chemistries you know.”]]></description><pubDate>Wed, 29 Apr 2026 12:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/rare-earth-free-magnets</guid><category>Permanent-magnets</category><category>Quantum-computers</category><category>Rare-earth-metals</category><category>Rare-earths</category><category>Electric-motors</category><dc:creator>Glenn Zorpette</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/photo-of-a-technician-working-on-a-cylindrical-machine-containing-brass-colored-components.jpg?id=66525633&amp;width=980"></media:content></item><item><title>Better Hardware Could Turn Zeros into AI Heroes</title><link>https://spectrum.ieee.org/sparse-ai</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/abstract-gradient-artwork-of-a-stylized-robot-head-with-circuits-and-binary-code-patterns.jpg?id=65862907&width=1245&height=700&coordinates=0%2C760%2C0%2C761"/><br/><br/><p><strong>When it comes to</strong> AI models, size matters.</p><p>Even though some artificial-intelligence experts <a href="https://spectrum.ieee.org/chain-of-thought-prompting" target="_self">warn</a> that scaling up large language models (LLMs) is hitting diminishing performance returns, companies are still coming out with ever larger AI tools. Meta’s latest Llama release had a staggering <a href="https://ai.meta.com/blog/llama-4-multimodal-intelligence/" rel="noopener noreferrer" target="_blank">2 trillion</a> parameters that define the model.</p><p>As models grow in size, their <a href="https://arxiv.org/abs/2001.08361" rel="noopener noreferrer" target="_blank">capabilities</a> increase. But so do the energy demands and the time it takes to run the models, which increases their <a href="https://spectrum.ieee.org/ai-index-2025" target="_self">carbon footprint</a>. To mitigate these issues, people have turned to <a href="https://spectrum.ieee.org/large-language-models-size" target="_self">smaller, less capable models</a> and using <a href="https://spectrum.ieee.org/1-bit-llm" target="_self">lower-precision</a> numbers whenever possible for the model parameters.</p><p>But there is another path that may retain a staggeringly large model’s high performance while reducing the time it takes to run an energy footprint. This approach involves befriending the zeros inside large AI models.</p><p>For many models, most of the parameters—the weights and activations—are actually zero, or so close to zero that they could be treated as such without losing accuracy. This quality is known as sparsity. Sparsity offers a significant opportunity for computational savings: Instead of wasting time and energy adding or multiplying zeros, these calculations could simply be skipped; rather than storing lots of zeros in memory, one need only store the nonzero parameters.</p><p>Unfortunately, today’s popular hardware, like multicore CPUs and GPUs, do not naturally take full advantage of sparsity. To fully leverage sparsity, researchers and engineers need to rethink and re-architect each piece of the design stack, including the hardware, low-level firmware, and application software.</p><p>In our research group at Stanford University, we have developed the first (to our knowledge) piece of hardware that’s capable of calculating all kinds of sparse and traditional workloads efficiently. The energy savings varied widely over the workloads, but on average our chip consumed one-seventieth the energy of a CPU, and performed the computation on average eight times as fast. To do this, we had to engineer the hardware, low-level firmware, and software from the ground up to take advantage of sparsity. We hope this is just the beginning of hardware and model development that will allow for more energy-efficient AI.</p><h2>What is sparsity?</h2><p>Neural networks, and the data that feeds into them, are represented as arrays of numbers. These arrays can be one-dimensional (vectors), two-dimensional (matrices), or more (tensors). A sparse vector, matrix, or tensor has mostly zero elements. The level of sparsity varies, but when zeroes make up more than 50 percent of any type of array, it can stand to benefit from sparsity-specific computational methods. In contrast, an object that is not sparse—that is, it has few zeros compared with the total number of elements—is called dense.</p><p>Sparsity can be naturally present, or it can be induced. For example, a <a href="https://arxiv.org/abs/2005.00687" rel="noopener noreferrer" target="_blank">social-network graph</a> will be naturally sparse. Imagine a graph where each node (point) represents a person, and each edge (a line segment connecting the points) represents a friendship. Since most people are not friends with one another, a matrix representing all possible edges will be mostly zeros. Other popular applications of AI, such as other forms of graph learning and <a href="https://arxiv.org/abs/1906.03109" rel="noopener noreferrer" target="_blank">recommendation models</a>, contain naturally occurring sparsity as well.</p><h3></h3><br/><img alt="Diagram mapping a sparse matrix to a fibertree and compressed storage format" class="rm-shortcode" data-rm-shortcode-id="d0cc84749a0f0fb374e27ea2ba2041c3" data-rm-shortcode-name="rebelmouse-image" id="3b584" loading="lazy" src="https://spectrum.ieee.org/media-library/diagram-mapping-a-sparse-matrix-to-a-fibertree-and-compressed-storage-format.jpg?id=65866445&width=980"/><h3></h3><br/><p>Beyond naturally occurring sparsity, sparsity can also be induced within an AI model in several ways. Two years ago, a team at <a href="https://spectrum.ieee.org/cerebras-wafer-scale-engine" target="_self">Cerebras</a> <a href="https://www.cerebras.ai/blog/introducing-sparse-llama-70-smaller-3x-faster-full-accuracy" target="_blank">showed</a> that one can set up to 70 to 80 percent of parameters in an LLM to zero without losing any accuracy. Cerebras demonstrated these results specifically on Meta’s open-source Llama 7B model, but the ideas extend to other LLM models like ChatGPT and Claude.</p><h2>The case for sparsity</h2><p>Sparse computation’s efficiency stems from two fundamental properties: the ability to compress away zeros and the convenient mathematical properties of zeros. Both the algorithms used in sparse computation and the hardware dedicated to them leverage these two basic ideas.</p><p>First, sparse data can be compressed, making it more memory efficient to store “sparsely”—that is, in something called a sparse data type. Compression also makes it more energy efficient to move data when dealing with large amounts of it. This is best understood by an example. Take a four-by-four matrix with three nonzero elements. Traditionally, this matrix would be stored in memory as is, taking up 16 spaces. This matrix can also be compressed into a sparse data type, getting rid of the zeros and saving only the nonzero elements. In our example, this results in 13 memory spaces as opposed to 16 for the dense, uncompressed version. These savings in memory increase with increased sparsity and matrix size.</p><h3></h3><br/><img alt="Diagram comparing dense and sparse matrix\u2013vector multiplication step by step." class="rm-shortcode" data-rm-shortcode-id="3d04f283be99eec83a4206f10d0394ca" data-rm-shortcode-name="rebelmouse-image" id="f523b" loading="lazy" src="https://spectrum.ieee.org/media-library/diagram-comparing-dense-and-sparse-matrix-u2013vector-multiplication-step-by-step.jpg?id=66499008&width=980"/><p><br/></p><p>In addition to the actual data values, compressed data also requires metadata. The row and column locations of the nonzero elements also must be stored. This is usually thought of as a “fibertree”: The row labels containing nonzero elements are listed and linked to the column labels of the nonzero elements, which are then linked to the values stored in those elements.</p><p>In memory, things get a bit more complicated still: The row and column labels for each nonzero value must be stored as well as the “segments” that indicate how many such labels to expect, so the metadata and data can be clearly delineated from one another.</p><p>In a dense, noncompressed matrix data type, values can be accessed either one at a time or in parallel, and their locations can be calculated directly with a simple equation. However, accessing values in sparse, compressed data requires looking up the coordinates of the row index and using that information to “indirectly” look up the coordinates of the column index before finally reaching the value. Depending on the actual locations of the sparse data values, these indirect lookups can be extremely random, making the computation data-dependent and requiring the allocation of memory lookups on the fly.</p><p>Second, two mathematical properties of zero let software and hardware skip a lot of computation. Multiplying any number by zero will result in a zero, so there’s no need to actually do the multiplication. Adding zero to any number will always return that number, so there’s no need to do the addition either.</p><p>In matrix-vector multiplication, one of the most common operations in AI workloads, all computations except those involving two nonzero elements can simply be skipped. Take, for example, the four-by-four matrix from the previous example and a vector of four numbers. In dense computation, each element of the vector must be multiplied by the corresponding element in each row and then added together to compute the final vector. In this case, that would take 16 multiplication operations and 16 additions (or four accumulations).</p><p>In sparse computation, only the nonzero elements of the vector need be considered. For each nonzero vector element, indirect lookup can be used to find any corresponding nonzero matrix element, and only those need to be multiplied and added. In the example shown here, only two multiplication steps will be performed, instead of 16.</p><h2>The trouble with GPUs and CPUs</h2><p>Unfortunately, modern hardware is not well suited to accelerating sparse computation. For example, say we want to perform a matrix-vector multiplication. In the simplest case, in a single CPU core, each element in the vector would be multiplied sequentially and then written to memory. This is slow, because we can do only one multiplication at a time. So instead people use CPUs with vector support or GPUs. With this hardware, all elements would be multiplied in parallel, greatly speeding up the application. Now, imagine that both the matrix and vector contain extremely sparse data. The vectorized CPU and GPU would spend most of their efforts multiplying by zero, performing completely ineffectual computations.</p><p><a href="https://developer.nvidia.com/blog/accelerating-inference-with-sparsity-using-ampere-and-tensorrt/" target="_blank">Newer generations</a> of GPUs are capable of taking some advantage of sparsity in their hardware, but only a particular kind, called structured sparsity. Structured sparsity assumes that two out of every four adjacent parameters are zero. However, some models benefit more from unstructured sparsity—the ability for any parameter (weight or activation) to be zero and compressed away, regardless of where it is and what it is adjacent to. GPUs can run unstructured sparse computation in software, for example, through the use of the <a href="https://docs.nvidia.com/cuda/cusparse/" target="_blank">cuSparse GPU library</a>. However, the support for sparse computations is often limited, and the GPU hardware gets underutilized, wasting energy-intensive computations on overhead.</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" rel="float: left;" style="float: left;"> <img alt="Neon pixel art of a glowing portal framed by geometric stairs and circuitry lines" class="rm-shortcode" data-rm-shortcode-id="7edb9085f930de797a7c401b9485d3ea" data-rm-shortcode-name="rebelmouse-image" id="012af" loading="lazy" src="https://spectrum.ieee.org/media-library/neon-pixel-art-of-a-glowing-portal-framed-by-geometric-stairs-and-circuitry-lines.jpg?id=65863062&width=980"/> <small class="image-media media-photo-credit" placeholder="Add Photo Credit..."><a href="https://petrapeterffy.com/" target="_blank">Petra Péterffy</a></small></p><p>When doing sparse computations in software, modern CPUs may be a better alternative to GPU computation, because they are designed to be more flexible. Yet, sparse computations on the CPU are often bottlenecked by the indirect lookups used to find nonzero data. CPUs are designed to “prefetch” data based on what they expect they’ll need from memory, but for randomly sparse data, that process often fails to pull in the right stuff from memory. When that happens, the CPU must waste cycles calling for the right data.</p><p>Apple was the <a href="https://ieeexplore.ieee.org/document/9833570" target="_blank">first</a> to speed up these indirect lookups by supporting a method called an array-of-pointers access pattern in the prefetcher of their A14 and M1 chips. Although innovations in prefetching make Apple CPUs more competitive for sparse computation, CPU architectures still have fundamental overheads that a dedicated sparse computing architecture would not, because they need to handle general-purpose computation.</p><p>Other companies have been developing <a href="https://spectrum.ieee.org/nvidia-ai" target="_self">hardware</a> that accelerates sparse machine learning as well. These include Cerebras’s <a href="https://spectrum.ieee.org/cerebras-chip-cs3" target="_self">Wafer Scale Engine</a> and <a href="https://ai.meta.com/blog/next-generation-meta-training-inference-accelerator-AI-MTIA/" target="_blank">Meta’s Training and Inference Accelerator (MTIA)</a>. The Wafer Scale Engine, and its corresponding sparse programming framework, have <a href="https://www.cerebras.ai/blog/introducing-sparse-llama-70-smaller-3x-faster-full-accuracy" target="_blank">shown</a> incredibly sparse results of up to 70 percent sparsity on LLMs. However, the company’s hardware and software solutions support only weight sparsity, not activation sparsity, which is important for many applications. The second version of the MTIA <a href="https://ai.meta.com/blog/next-generation-meta-training-inference-accelerator-AI-MTIA/" target="_blank">claims</a> a sevenfold sparse compute performance boost over the <a href="https://doi-org.stanford.idm.oclc.org/10.1145/3579371.3589348" target="_blank">MTIA v1</a>. However, the only publicly available information regarding sparsity support in the MTIA v2 is for matrix multiplication, not for vectors or tensors.</p><p>Although matrix multiplications take up the majority of computation time in most modern ML models, it’s important to have sparsity support for other parts of the process. To avoid switching back and forth between sparse and dense data types, all of the operations should be sparse.</p><h2>Onyx</h2><p>Instead of these halfway solutions, our team at Stanford has developed a hardware accelerator, <a href="https://ieeexplore.ieee.org/document/10631383" target="_blank">Onyx</a>, that can take advantage of sparsity from the ground up, whether it’s structured or unstructured. Onyx is the first programmable accelerator to support both sparse and dense computation; it’s capable of accelerating key operations in both domains.</p>To understand Onyx, it is useful to know what a coarse-grained reconfigurable array (CGRA) is and how it compares with more familiar hardware, like CPUs and field-programmable gate arrays (FPGAs).<p>CPUs, CGRAs, and FPGAs represent a trade-off between efficiency and flexibility. Each individual logic unit of a CPU is designed for a specific function that it performs efficiently. On the other hand, since each individual bit of an FPGA is configurable, these arrays are extremely flexible, but very inefficient. The goal of CGRAs is to achieve the flexibility of FPGAs with the efficiency of CPUs.</p><p>CGRAs are composed of efficient and configurable units, typically memory and compute, that are specialized for a particular application domain. This is the key benefit of this type of array: Programmers can reconfigure the internals of a CGRA at a high level, making it more efficient than an FPGA but more flexible than a CPU.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Two circuit boards and a pen showing a chip shrinking from large to tiny size." class="rm-shortcode" data-rm-shortcode-id="b8111010f181900745167f0ffb5617f3" data-rm-shortcode-name="rebelmouse-image" id="f394d" loading="lazy" src="https://spectrum.ieee.org/media-library/two-circuit-boards-and-a-pen-showing-a-chip-shrinking-from-large-to-tiny-size.jpg?id=65970072&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">The Onyx chip, built on a coarse-grained reconfigurable array (CGRA), is the first (to our knowledge) to support both sparse and dense computations. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Olivia Hsu</small></p><p>Onyx is composed of flexible, programmable processing element (PE) tiles and memory (MEM) tiles. The memory tiles store compressed matrices and other data formats. The processing element tiles operate on compressed matrices, eliminating all unnecessary and ineffectual computation.</p><p>The Onyx compiler handles conversion from software instructions to CGRA configuration. First, the input expression—for instance, a sparse vector multiplication—is translated into a graph of abstract memory and compute nodes. In this example, there are memories for the input vectors and output vectors, a compute node for finding the intersection between nonzero elements, and a compute node for the multiplication. The compiler figures out how to map the abstract memory and compute nodes onto MEMs and PEs on the CGRA, and then how to route them together so that they can transfer data between them. Finally, the compiler produces the instruction set needed to configure the CGRA for the desired purpose.</p><p>Since Onyx is programmable, engineers can map many different operations, such as vector-vector element multiplication, or the key tasks in AI, like matrix-vector or matrix-matrix multiplication, onto the accelerator.</p><p>We evaluated the efficiency gains of our hardware by looking at the product of energy used and the time it took to compute, called the energy-delay product (EDP). This metric captures the trade-off of speed and energy. Minimizing just energy would lead to very slow devices, and minimizing speed would lead to high-area, high-power devices.</p><p>Onyx achieves up to 565 times as much energy-delay product over CPUs (we used a 12-core Intel Xeon CPU) that utilize dedicated sparse libraries. Onyx can also be configured to accelerate regular, dense applications, similar to the way a GPU or TPU would. If the computation is sparse, Onyx is configured to use sparse primitives, and if the computation is dense, Onyx is reconfigured to take advantage of parallelism, similar to how GPUs function. This architecture is a step toward a single system that can accelerate both sparse and dense computations on the same silicon.</p><p>Just as important, Onyx enables new algorithmic thinking. Sparse acceleration hardware will not only make AI more performance- and energy efficient but also enable researchers and engineers to explore new algorithms that have the potential to dramatically improve AI.</p><h2>The future with sparsity</h2><p>Our team is already working on next-generation chips built off of Onyx. Beyond matrix multiplication operations, machine learning models perform other types of math, like nonlinear layers, normalization, the softmax function, and more. We are adding support for the full range of computations on our next-gen accelerator and within the compiler. Since sparse machine learning models may have both sparse and dense layers, we are also working on integrating the dense and sparse accelerator architecture more efficiently on the chip, allowing for fast transformation between the different data types. We’re also looking at ways to manage memory constraints by breaking up the sparse data more effectively so we can run computations on several sparse accelerator chips.</p><p>We are also working on systems that can predict the performance of accelerators such as ours, which will help in designing better hardware for sparse AI. Longer term, we’re interested in seeing whether high degrees of sparsity throughout AI computation will catch on with more model types, and whether sparse accelerators become adopted at a larger scale.</p><p>Building the hardware to unstructured sparsity and optimally take advantage of zeros is just the beginning. With this hardware in hand, AI researchers and engineers will have the opportunity to explore new models and algorithms that leverage sparsity in novel and creative ways. We see this as a crucial research area for managing the ever-increasing runtime, costs, and environmental impact of AI. <span class="ieee-end-mark"></span></p>]]></description><pubDate>Tue, 28 Apr 2026 18:03:40 +0000</pubDate><guid>https://spectrum.ieee.org/sparse-ai</guid><category>Ai-models</category><category>Gpus</category><category>Energy-efficiency</category><category>Data-compression</category><dc:creator>Olivia Hsu</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/abstract-gradient-artwork-of-a-stylized-robot-head-with-circuits-and-binary-code-patterns.jpg?id=65862907&amp;width=980"></media:content></item><item><title>The Chip That Made Hardware Rewriteable</title><link>https://spectrum.ieee.org/fpga-chip-ieee-milestone</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/die-photo-of-an-integrated-circuit-with-an-8-by-8-array.jpg?id=66633028&width=1245&height=700&coordinates=0%2C469%2C0%2C469"/><br/><br/><p>Many of the world’s most advanced electronic systems—including <a href="https://spectrum.ieee.org/tag/routers" target="_self">Internet routers</a>, <a href="https://spectrum.ieee.org/6g-wireless" target="_self">wireless base stations</a>, <a href="https://spectrum.ieee.org/mri" target="_self">medical imaging scanners</a>, and <a href="https://www.ibm.com/think/topics/ai-accelerator" rel="noopener noreferrer" target="_blank">some artificial intelligence tools</a>—depend on <a href="https://spectrum.ieee.org/tag/fpga" target="_self">field-programmable gate arrays</a>. Computer chips with internal hardware circuits, the FPGAs can be reconfigured after manufacturing.</p><p>On 12 March, an <a href="https://ieeemilestones.ethw.org/Main_Page" rel="noopener noreferrer" target="_blank">IEEE Milestone</a> plaque recognizing the first FPGA was dedicated at the <a href="https://www.amd.com/en.html" rel="noopener noreferrer" target="_blank">Advanced Micro Devices</a> campus in San Jose, Calif., the former <a href="https://en.wikipedia.org/wiki/Xilinx" rel="noopener noreferrer" target="_blank">Xilinx</a> headquarters and the birthplace of the technology.</p><p>The FPGA earned the Milestone designation because it introduced iteration to semiconductor design. Engineers could redesign hardware repeatedly without fabricating a new chip, dramatically reducing development risk and enabling faster innovation at a time when semiconductor costs were rising rapidly.</p><p>The ceremony, which was organized by the <a href="https://ieeescv.org/" rel="noopener noreferrer" target="_blank">IEEE Santa Clara Valley Section</a>, brought together professionals from across the semiconductor industry and IEEE leadership. Speakers at the event included <a href="https://www.seti.org/people/stephen-trimberger/" rel="noopener noreferrer" target="_blank">Stephen Trimberger</a>, an IEEE and <a href="https://www.acm.org/" rel="noopener noreferrer" target="_blank">ACM</a> Fellow<a href="https://www.seti.org/people/stephen-trimberger/" rel="noopener noreferrer" target="_blank"></a>whose technical contributions helped shape modern FPGA architecture. Trimberger reflected on how the invention enabled software-programmable hardware.</p><h2>Solving computing’s flexibility-performance tradeoff</h2><p>FPGAs emerged in the 1980s to address a core limitation in computing. A microprocessor executes software instructions sequentially, making it flexible but sometimes too slow for workloads requiring many operations at once.</p><p>At the other extreme, <a href="https://spectrum.ieee.org/lowbudget-chip-design-how-hard-is-it" target="_self">application-specific integrated circuits</a> are chips designed to do only one task. ASICs achieve high efficiency but require lengthy development cycles and nonrecurring engineering costs, which are large, upfront investments. Expenses include designing the chip and preparing it for manufacturing—a process that involves creating detailed layouts, building <a href="https://spectrum.ieee.org/leading-chipmakers-eye-euv-lithography-to-save-moores-law" target="_self">masks for the fabrication machines</a>, and setting up production lines to handle the tiny circuits.</p><p>“ASICs can deliver the best performance, but the development cycle is long and the nonrecurring engineering cost can be very high,” says <a href="https://vast.cs.ucla.edu/people/faculty/jason-cong" rel="noopener noreferrer" target="_blank">Jason Cong</a>, an IEEE Fellow and professor of computer science at the <a href="https://samueli.ucla.edu/" rel="noopener noreferrer" target="_blank">University of California, Los Angeles</a>. “FPGAs provide a sweet spot between processors and custom silicon.”</p><p>Cong’s foundational work in FPGA design automation and high-level synthesis transformed how reconfigurable systems are programmed. He developed synthesis tools that translate <a href="https://spectrum.ieee.org/top-programming-languages-2025" target="_self">C/C++</a> into hardware designs, for example.</p><p>At the heart of his work is an underlying principle first espoused by electrical engineer <a href="https://www.invent.org/inductees/ross-freeman" rel="noopener noreferrer" target="_blank">Ross Freeman</a>: By configuring hardware using programmable memory embedded inside the chip, FPGAs combine hardware-level speed with the adaptability traditionally associated with software.</p><h2>Silicon Valley origins: the first FPGA</h2><p>The FPGA architecture originated in the mid-1980s at Xilinx, a Silicon Valley company founded in 1984. The invention is widely credited to Freeman, a Xilinx cofounder and the startup’s CTO. He envisioned a chip with circuitry that could be configured after fabrication rather than fixed permanently during creation.</p><p>Articles about the <a href="https://www.eejournal.com/article/how-the-fpga-came-to-be-part-5/" rel="noopener noreferrer" target="_blank">history of the FPGA</a> emphasize that he saw it as a deliberate break from conventional chip design.</p><p>At the time, semiconductor engineers treated <a href="https://spectrum.ieee.org/special-reports/the-transistor-at-75/" target="_self">transistors</a> as scarce resources. Custom chips were carefully optimized so that nearly every transistor served a specific purpose.</p><p>Freeman proposed a different approach. He figured <a href="https://spectrum.ieee.org/special-reports/50-years-of-moores-law/" target="_self">Moore’s Law</a> would soon change chip economics. The principle holds that transistor counts roughly double every two years, making computing cheaper and more powerful. Freeman posited that as transistors became abundant, flexibility would matter more than perfect efficiency.</p><p>He envisioned a device composed of programmable logic blocks connected through configurable routing—a chip filled with what he described as “open gates,” ready to be defined by users after manufacturing. Instead of fixing hardware in silicon permanently, engineers could configure and reconfigure circuits as requirements evolved.</p><p>Freeman sometimes compared the concept to a blank cassette tape: Manufacturers would supply the medium, while engineers determined its function. The analogy captured a profound shift in who controls the technology, shifting hardware design flexibility from chip fabrication facilities to the system designers themselves.</p><p>In 1985 Xilinx introduced the first FPGA for commercial sale: the <a href="https://spectrum.ieee.org/chip-hall-of-fame-xilinx-xc2064-fpga" target="_self">XC2064</a>. The device contained 64 configurable logic blocks—small digital circuits capable of performing logical operations—arranged in an 8-by-8 grid. Programmable routing channels allowed engineers to define how signals moved between blocks, effectively wiring a custom circuit with software.</p><p>Fabricated using a 2-micrometer process (meaning that 2 µm was the minimum size of the features that could be patterned onto silicon using <a href="https://www.micron.com/content/dam/micron/educatorhub/fabrication/photolithography/micron-fabrication-intro-to-photolithography-presentation.pdf" rel="noopener noreferrer" target="_blank">photolithography</a>), the XC2064 implemented a few thousand logic gates. Modern FPGAs can contain hundreds of millions of gates, enabling vastly more complex designs. Yet the XC2064 established a design workflow still used today: Engineers describe the hardware behavior digitally and then “compile the design,” a process that automatically translates the plans into the instructions the FPGA needs to set its logic blocks and wiring, according to <a href="https://www.amd.com/en.html" rel="noopener noreferrer" target="_blank">AMD</a>. Engineers then load that configuration onto the chip.</p><h2>The breakthrough: hardware defined by memory</h2><p>Earlier <a href="https://tessellatedcircuits.com/pld_hist.php" rel="noopener noreferrer" target="_blank">programmable logic devices</a>, such as erasable programmable read-only memory, or EPROM, allowed limited customization but relied on largely fixed wiring structures that <a href="https://medium.com/@najamhassan569/understanding-plds-the-building-blocks-of-modern-digital-systems-dbefd69fbc21" rel="noopener noreferrer" target="_blank">did not scale well</a> as circuits grew more complex, Cong says.</p><p>FPGAs introduced programmable interconnects—networks of electronic switches controlled by memory cells distributed across the chip. When powered on, the device loads a <a href="https://spectrum.ieee.org/computing-with-random-pulses-promises-to-simplify-circuitry-and-save-power" target="_self">bitstream</a> configuration file that determines how its internal circuits behave.</p><p>“As process technology improved and transistor counts increased, the cost of programmability became much less significant,” Cong says.</p><h2>From “glue logic” to essential infrastructure</h2><p>“Initially, FPGAs were used as what engineers called <a href="https://www.pcmag.com/encyclopedia/term/glue-logic" rel="noopener noreferrer" target="_blank">glue logic</a>,” Cong says.</p><p><em><em>Glue logic</em></em> refers to simple circuits that connect processors, memory, and peripheral devices so the system works reliably, according to <a href="https://www.pcmag.com/encyclopedia/term/glue-logic" rel="noopener noreferrer" target="_blank"><em><em>PC Magazine</em></em></a>. In other words, it “glues” different components together, especially when interfaces change frequently.</p><p>Early adopters recognized the advantage of hardware that could adapt as standards evolved. In “<a href="https://cacm.acm.org/practice/the-history-status-and-future-of-fpgas/" rel="noopener noreferrer" target="_blank">The History, Status, and Future of FPGAs</a>,” published in <a href="https://cacm.acm.org/" rel="noopener noreferrer" target="_blank"><em><em>Communications of the ACM</em></em></a>, engineers at Xilinx and organizations such as <a href="https://spectrum.ieee.org/7-bell-labs-ieee-milestones" target="_self">Bell Labs</a>, <a href="https://computerhistory.org/blog/fairchild-semiconductor-the-60th-anniversary-of-a-silicon-valley-legend/" rel="noopener noreferrer" target="_blank">Fairchild Semiconductor</a>, <a href="https://www.ibm.com/us-en" rel="noopener noreferrer" target="_blank">IBM</a>, and <a href="https://www.britannica.com/money/Sun-Microsystems-Inc" rel="noopener noreferrer" target="_blank">Sun Microsystems</a> said the earliest uses of <a href="https://www.eetimes.com/transfer-from-fpgas-for-prototype-to-asics-for-production/" rel="noopener noreferrer" target="_blank">FPGAs were for prototyping ASICs</a>. They also used it for <a href="https://www.synopsys.com/glossary/what-is-hav-emulation.html" rel="noopener noreferrer" target="_blank">validating complex systems</a> by running their software before fabrication, allowing the companies to deploy specialized products manufactured in modest volumes.</p><p>Those uses revealed a broader shift: Hardware no longer needed to remain fixed once deployed.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A group dressed in business casual attire smiling and posing together around an outdoor bench adorned with a plaque." class="rm-shortcode" data-rm-shortcode-id="d28b2fa5d3ac1b68dd9ced85e46da61a" data-rm-shortcode-name="rebelmouse-image" id="c3363" loading="lazy" src="https://spectrum.ieee.org/media-library/a-group-dressed-in-business-casual-attire-smiling-and-posing-together-around-an-outdoor-bench-adorned-with-a-plaque.jpg?id=66633157&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">Attendees at the Milestone plaque dedication ceremony included (seated L to R) 2025 IEEE President Kathleen Kramer, 2024 IEEE President Tom Coughlin, and Santa Clara Valley Section Milestones Chair Brian Berg.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Douglas Peck/AMD</small></p><h2>Semiconductor economics changed the equation</h2><p>The rise of FPGAs closely followed changes in semiconductor economics, Cong says.</p><p>Developing a custom chip requires a large upfront investment before production begins. As fabrication costs increased, products had to ship in large quantities to make ASIC development economically viable, according to <a href="https://anysilicon.com/the-economics-of-asic/" target="_blank">a post</a> published by <a href="https://anysilicon.com/" target="_blank">AnySilicon</a>.</p><p>FPGAs allowed designers to move forward without that larger monetary commitment.</p><p>ASIC development typically requires 18 to 24 months from conception to silicon, while FPGA implementations often can be completed within three to six months using modern design tools, Cong says. The shorter cycle and the ability to reconfigure the hardware enabled startups, universities, and equipment manufacturers to experiment with advanced architectures that were previously accessible mainly to large chip companies.</p><h2>Lookup tables and the rise of reconfigurable computing</h2><p>A popular technique for implementing mathematical functions in hardware is<a href="https://ieeexplore.ieee.org/document/10013797" target="_blank"></a>the <a href="https://ieeexplore.ieee.org/document/10013797" rel="noopener noreferrer" target="_blank">lookup table</a> (LUT). A LUT is a small memory element that stores the results of logical operations, according to “<a href="https://arxiv.org/abs/2511.06174" rel="noopener noreferrer" target="_blank">LUT-LLM: Efficient Large Language Model Inference with Memory-based Computations on FPGAs</a>,” a paper selected for presentation next month at the 34th <a href="https://www.fccm.org/" rel="noopener noreferrer" target="_blank">IEEE International Symposium on Field-Programmable Custom Computing Machines</a> (FCCM).</p><p>Instead of repeatedly recalculating outcomes, the chip retrieves answers directly from memory. Cong compares the approach to consulting multiplication tables rather than recomputing the arithmetic each time.</p><p>Research led by Cong and others helped develop efficient methods for mapping digital circuits onto LUT-based architectures, shaping routing and layout strategies used in modern devices.</p><p>As transistor budgets expanded, FPGA vendors integrated memory blocks, digital signal-processing units, high-speed communication interfaces, <a href="https://spectrum.ieee.org/tag/cryptography" target="_self">cryptographic engines</a>, and embedded processors, transforming the devices into versatile computing platforms.</p><h2>Why the gate arrays are distinct from CPUs, GPUs, and ASICs</h2><p>FPGAs coexist with other processors because each one optimizes different priorities. Central processing units excel at general computing. Graphics processing units, designed to perform many calculations simultaneously, dominate large parallel workloads such as AI training. ASICs provide maximum efficiency when designs remain stable and production volumes are high.</p><p class="pull-quote">“ASICs can deliver the best performance, but the development cycle is long, and the nonrecurring engineering cost can be very high. FPGAs provide a sweet spot between processors and custom silicon.” <strong>—Jason Cong, IEEE Fellow and professor of computer science at UCLA.</strong></p><p>“FPGAs are not replacements for CPUs or GPUs,” Cong says. “They complement those processors in <a href="https://docs.lib.purdue.edu/cgi/viewcontent.cgi?article=1209&context=ecetr" target="_blank">heterogeneous computing</a> systems.”</p><p>Modern computing platforms increasingly combine multiple types of processors to balance flexibility, performance, and energy efficiency.</p><h2>A Milestone for an idea, not just a device</h2><p>This IEEE Milestone recognizes more than a successful semiconductor product. It also acknowledges a shift in how engineers innovate.</p><p>Reconfigurable hardware allows designers to test ideas quickly, refine architectures, and deploy systems while standards and markets evolve.</p><p>“Without FPGAs,” Cong says, “the pace of hardware innovation would likely be much slower.”</p><p>Four decades after the first FPGA appeared, the technology’s enduring legacy reflects Freeman’s insight: Hardware did not need to remain fixed. By accepting a small amount of unused silicon in exchange for adaptability, engineers transformed chips from static products into platforms for continuous experimentation—turning silicon itself into a medium engineers could rewrite.</p><p>Among those who attended the Milestone ceremony were 2025 IEEE President <a href="https://www.linkedin.com/in/kathleenkramer" target="_blank">Kathleen Kramer</a>; 2024 IEEE President <a href="https://corporate-awards.ieee.org/speaker/tom-coughlin/" rel="noopener noreferrer" target="_blank">Tom Coughlin</a>; <a href="https://www.linkedin.com/in/averylu" rel="noopener noreferrer" target="_blank">Avery Lu</a>, chair of the <a href="https://ieeescv.org/" rel="noopener noreferrer" target="_blank">IEEE Santa Clara Valley Section</a>; and <a href="https://ieeetv.ieee.org/speaker/brian-berg" rel="noopener noreferrer" target="_blank">Brian Berg</a>, history and milestones chair of <a href="https://ieee-region6.org/" rel="noopener noreferrer" target="_blank">IEEE Region 6</a><a href="https://ieeetv.ieee.org/speaker/brian-berg" rel="noopener noreferrer" target="_blank">. They joined</a> AMD’s chief executive, <a href="https://www.amd.com/en/corporate/leadership/lisa-su.html" rel="noopener noreferrer" target="_blank">Lisa Su</a>, and <a href="https://www.amd.com/en/corporate/leadership/salil-raje.html#:~:text=Salil%20Raje%20is%20senior%20vice,with%20an%20emphasis%20on%20growing" rel="noopener noreferrer" target="_blank">Salil Raje</a>, senior vice president and general manager of adaptive and embedded computing at AMD.</p><p>The <a href="https://ethw.org/Milestones:Field_Programmable_Gate_Array" rel="noopener noreferrer" target="_blank">IEEE Milestone plaque</a> honoring the field-programmable gate array reads:</p><p><em><em>“</em></em><em><em>The FPGA is an integrated circuit with user-programmable Boolean logic functions and interconnects. FPGA inventor Ross Freeman cofounded Xilinx to productize his 1984 invention, and in 1985 the XC2064 was introduced with 64 programmable 4-input logic functions. Xilinx’s FPGAs helped accelerate a dramatic industry shift wherein ‘fabless’ companies could use software tools to design hardware while engaging ‘foundry’ companies to handle the capital-intensive task of manufacturing the software-defined hardware.”</em></em></p><p>Administered by the <a href="https://www.ieee.org/about/history-center?check_logged_in=1" rel="noopener noreferrer" target="_blank">IEEE History Center</a> and supported by donors, the IEEE Milestone program recognizes outstanding technical developments worldwide that are at least 25 years old.</p><p>Check out <em><em>Spectrum</em></em>’s <a href="https://connect.ieee.org/NzU2LUdQSC04OTkAAAGhT7-QweL2i3BmX2b-_PBdiukfOVwCR2UPcYg1G4khUu5odaR3T07IAVEY5ylL-hWj7LNbRKU=" rel="noopener noreferrer" target="_blank">History of Technology</a> channel to read more stories about key engineering achievements.</p>]]></description><pubDate>Tue, 28 Apr 2026 18:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/fpga-chip-ieee-milestone</guid><category>Ieee-history</category><category>Fpga</category><category>Xilinx</category><category>Ieee-milestone</category><category>Amd</category><category>History-of-technology</category><category>Type-ti</category><dc:creator>Willie D. Jones</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/die-photo-of-an-integrated-circuit-with-an-8-by-8-array.jpg?id=66633028&amp;width=980"></media:content></item><item><title>GPU Renters Are Playing a Silicon Lottery</title><link>https://spectrum.ieee.org/gpu-performance-comparison</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/bar-chart-comparing-tesla-t4-a10g-a100-l4-and-h100-gpu-performance-ranges.png?id=65814435&width=980"/><br/><br/><p>Think one GPU is very much like another? Think again. It turns out that there’s surprising variability in the performance delivered by chips of the same model. That can make getting your money’s worth by renting time on a GPU from a cloud provider a real roll of the dice, according to research from the College of William & Mary, Jefferson Lab, and <a href="https://www.silicondata.com/" rel="noopener noreferrer" target="_blank">Silicon Data</a>.</p><p>“It’s called the silicon lottery,” says <a href="https://www.linkedin.com/in/carmenrli/" rel="noopener noreferrer" target="_blank">Carmen Li,</a> founder and CEO of Silicon Data, which tracks <a href="https://spectrum.ieee.org/gpu-prices" target="_self">GPU rental prices</a> and <a href="https://spectrum.ieee.org/mlperf-trends" target="_self">benchmarks</a> cloud-computing performance.</p><p>The <a href="https://www.computer.org/csdl/proceedings-article/sc/2022/544400a937/1I0bT7vc6B2" rel="noopener noreferrer" target="_blank">silicon lottery’s existence</a> has been known since at least 2022, when researchers at the University of Wisconsin tied it to variations in the performance of GPU-dependent supercomputers. Li and her colleagues figured that the effect would be even more pronounced for AI cloud customers.</p><h3>Performance varies for GPU models in the cloud</h3><br/><img alt="Chart comparing GPU models by 16-bit TFLOPS and median hourly rental prices." class="rm-shortcode" data-rm-shortcode-id="14114673d2c672cde525bd4d147097b7" data-rm-shortcode-name="rebelmouse-image" id="b5d4e" loading="lazy" src="https://spectrum.ieee.org/media-library/chart-comparing-gpu-models-by-16-bit-tflops-and-median-hourly-rental-prices.png?id=65816885&width=980"/><h3></h3><br/><p>So they ran 6,800 instances of the index firm’s benchmark test on 3,500 randomly selected GPUs operated by 11 cloud-computing providers. The 3,500 GPUs comprised <a href="https://en.wikipedia.org/wiki/List_of_Nvidia_graphics_processing_units" target="_blank">11 models of Nvidia GPU</a>, the most advanced being the <a href="https://spectrum.ieee.org/ai-benchmark-mlperf-llama-stablediffusion" target="_self">Nvidia H200</a> SXM. (The team wasn’t just picking on <a href="https://www.nvidia.com/en-us/" target="_blank">Nvidia</a>; the GPU giant makes up most of the rental cloud market.)</p><p>The benchmark, called <a href="https://www.silicondata.com/products/silicon-mark" target="_blank">SiliconMark</a>, is intended to provide a snapshot of a GPU’s ability to run large language models, or LLMs. It tests 16-bit floating-point computing performance, measured in trillions of operations per second, and a GPU’s internal-memory bandwidth, measured in gigabytes per second. <a href="https://downloads.silicondata.com/documents/GPGPU26_SiliconData.pdf" rel="noopener noreferrer" target="_blank">The results</a> showed that the computing performance varied for all models, but for the 259 H100 PCIe GPUs it differed by as much as 34.5 percent, and the memory bandwidth of the 253 H200 SXM GPUs varied by as much as 38 percent.</p><h3></h3><br/><img alt="Chart comparing GPU internal memory bandwidth by model, from Tesla T4 to H200 SXM." class="rm-shortcode" data-rm-shortcode-id="b5cdb54f4666983523d50b7fc5968cbe" data-rm-shortcode-name="rebelmouse-image" id="b818b" loading="lazy" src="https://spectrum.ieee.org/media-library/chart-comparing-gpu-internal-memory-bandwidth-by-model-from-tesla-t4-to-h200-sxm.png?id=65816932&width=980"/><p><span>Differences in how the GPU is cooled, how cloud operators configure their computers, and how much use the chip has seen can all contribute to variations in performance of otherwise identical chips. But Silicon Data’s analysis showed that the real culprit was variations in the chips themselves, likely due to manufacturing issues.</span></p><p>Such randomness has real dollars-and-cents consequences, the researchers argue, because there’s a chance that a pricier, more advanced GPU won’t deliver better performance than an older model chip.</p><p>So what should GPU renters do? “The most practical approach is to benchmark the actual rental they receive,” says <a href="https://www.linkedin.com/in/jcornick/" target="_blank">Jason Cornick</a>, head of infrastructure at Silicon Data. “Running a benchmark tool [such as SiliconMark] allows them to compare their specific instance’s performance against a broader corpus of data.”</p>]]></description><pubDate>Thu, 23 Apr 2026 18:06:01 +0000</pubDate><guid>https://spectrum.ieee.org/gpu-performance-comparison</guid><category>Artificial-intelligence</category><category>Cloud-computing</category><category>Nvidia</category><category>Gpus</category><category>Gpu</category><category>Hyperscalers</category><category>Graphics-processing-units</category><category>Benchmarking</category><category>Large-language-models</category><dc:creator>Samuel K. Moore</dc:creator><media:content medium="image" type="image/png" url="https://assets.rbl.ms/65814435/origin.png"></media:content></item><item><title>What Anthropic’s Mythos Means for the Future of Cybersecurity</title><link>https://spectrum.ieee.org/ai-cybersecurity-mythos</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-cgi-image-of-a-translucent-padlock-filled-with-0s-and-1s-one-spot-is-broken-and-the-numbers-are-spraying-out-of-that-spot.jpg?id=65714765&width=1245&height=700&coordinates=0%2C156%2C0%2C157"/><br/><br/><p>Two weeks ago, Anthropic <a href="https://red.anthropic.com/2026/mythos-preview/" rel="noopener noreferrer" target="_blank">announced</a> that its new model, Claude Mythos Preview, can autonomously find and weaponize software vulnerabilities, turning them into working exploits without expert guidance. These were vulnerabilities in key software like operating systems and internet infrastructure that thousands of software developers working on those systems failed to find. This capability will have major security implications, compromising the devices and services we use every day. As a result, <a href="https://spectrum.ieee.org/tag/anthropic" target="_blank">Anthropic</a> is not releasing the model to the general public, but instead to a <a href="https://www.anthropic.com/glasswing" rel="noopener noreferrer" target="_blank">limited number</a> of companies.</p><div class="rm-embed embed-media"><iframe height="110px" id="noa-web-audio-player" src="https://embed-player.newsoveraudio.com/v4?key=q5m19e&id=https://spectrum.ieee.org/ai-cybersecurity-mythos&bgColor=F5F5F5&color=1b1b1c&playColor=1b1b1c&progressBgColor=F5F5F5&progressBorderColor=bdbbbb&titleColor=1b1b1c&timeColor=1b1b1c&speedColor=1b1b1c&noaLinkColor=556B7D&noaLinkHighlightColor=FF4B00&feedbackButton=true" style="border: none" width="100%"></iframe></div><p><span>The news rocked the internet security community. There were few details in Anthropic’s announcement, </span><a href="https://srinstitute.utoronto.ca/news/the-mythos-question-who-decides-when-ai-is-too-dangerous" target="_blank">angering</a><span> many observers. Some speculate that Anthropic </span><a href="https://kingy.ai/ai/too-dangerous-to-release-or-just-too-expensive-the-real-reason-anthropic-is-hiding-its-most-powerful-ai/" target="_blank">doesn’t have</a><span> the GPUs to run the thing, and that cybersecurity was the excuse to limit its release. Others argue Anthropic is holding to its AI safety mission. </span><a href="https://www.nytimes.com/2026/04/07/opinion/anthropic-ai-claude-mythos.html" target="_blank">There’s</a><span> </span><a href="https://www.axios.com/2026/04/08/anthropic-mythos-model-ai-cyberattack-warning" target="_blank">hype</a><span> and </span><a href="https://www.artificialintelligencemadesimple.com/p/anthropics-claude-mythos-launch-is" target="_blank">counter</a><a href="https://aisle.com/blog/ai-cybersecurity-after-mythos-the-jagged-frontier" target="_blank">hype</a><span>, </span><a href="https://www.aisi.gov.uk/blog/our-evaluation-of-claude-mythos-previews-cyber-capabilities" target="_blank">reality</a><span> and marketing. It’s a lot to sort out, even if you’re an expert.</span></p><p>We see Mythos as a real but incremental step, one in a long line of incremental steps. But even incremental steps can be important when we look at the big picture.</p><h2>How AI Is Changing Cybersecurity</h2><p>We’ve <a href="https://spectrum.ieee.org/online-privacy" target="_self">written about</a> shifting baseline syndrome, a phenomenon that leads people—the public and experts alike—to discount massive long-term changes that are hidden in incremental steps. It has happened with online privacy, and it’s happening with AI. Even if the vulnerabilities found by Mythos could have been found using AI models from last month or last year, they couldn’t have been found by AI models from five years ago.</p><p>The Mythos announcement reminds us that AI has come a long way in just a few years: The baseline really has shifted. Finding vulnerabilities in source code is the type of task that today’s large language models excel at. Regardless of whether it happened last year or will happen next year, it’s been clear for a <a href="https://sockpuppet.org/blog/2026/03/30/vulnerability-research-is-cooked/" target="_blank">while</a> this kind of capability was coming soon. The question is how we <a href="https://labs.cloudsecurityalliance.org/mythos-ciso/" target="_blank">adapt to it</a>.</p><p>We don’t believe that an AI that can hack autonomously will create permanent asymmetry between offense and defense; it’s likely to be more <a href="https://danielmiessler.com/blog/will-ai-help-moreattackers-defenders" rel="noopener noreferrer" target="_blank">nuanced</a> than that. Some vulnerabilities can be found, verified, and patched automatically. Some vulnerabilities will be hard to find but easy to verify and patch—consider generic cloud-hosted web applications built on standard software stacks, where updates can be deployed quickly. Still others will be easy to find (even without powerful AI) and relatively easy to verify, but harder or impossible to patch, such as IoT appliances and industrial equipment that are rarely updated or can’t be easily modified.</p><p>Then there are systems whose vulnerabilities will be easy to find in code but difficult to verify in practice. For example, complex distributed systems and cloud platforms can be composed of thousands of interacting services running in parallel, making it difficult to distinguish real vulnerabilities from false positives and to reliably reproduce them.</p><p>So we must separate the patchable from the unpatchable, and the easy to verify from the hard to verify. This taxonomy also provides us guidance for how to protect such systems in an era of powerful AI vulnerability-finding tools.</p><p>Unpatchable or hard to verify systems should be protected by wrapping them in more restrictive, tightly controlled layers. You want your fridge or thermostat or industrial control system behind a restrictive and constantly updated firewall, not freely talking to the internet.</p><p>Distributed systems that are fundamentally interconnected should be traceable and should follow the principle of least privilege, where each component has only the access it needs. These are bog-standard security ideas that we might have been tempted to throw out in the era of AI, but they’re still as relevant as ever.</p><h2>Rethinking Software Security Practices</h2><p>This also raises the salience of best practices in software engineering. Automated, thorough, and continuous testing was always important. Now we can take this practice a step further and use defensive AI agents to <a href="https://www.secwest.net/ai-triage" rel="noopener noreferrer" target="_blank">test exploits</a> against a real stack, over and over, until the false positives have been weeded out and the real vulnerabilities and fixes are confirmed. This kind of <a href="https://www.csoonline.com/article/4069075/autonomous-ai-hacking-and-the-future-of-cybersecurity.html" rel="noopener noreferrer" target="_blank">VulnOps</a> is likely to become a standard part of the development process.</p><p>Documentation becomes more valuable, as it can guide an AI agent on a bug-finding mission just as it does developers. And following standard practices and using standard tools and libraries allows AI and engineers alike to recognize patterns more effectively, even in a world of individual and ephemeral <a href="https://www.csoonline.com/article/4152133/cybersecurity-in-the-age-of-instant-software.html" rel="noopener noreferrer" target="_blank">instant software</a>—code that can be generated and deployed on demand.</p><p>Will this favor <a href="https://www.schneier.com/essays/archives/2018/03/artificial_intellige.html" rel="noopener noreferrer" target="_blank">offense or defense</a>? The defense eventually, probably, especially in systems that are easy to patch and verify. Fortunately, that includes our phones, web browsers, and major internet services. But today’s cars, electrical transformers, fridges, and lampposts are connected to the internet. Legacy banking and airline systems are networked.</p>Not all of those are going to get patched as fast as needed, and we may see a few years of constant hacks until we arrive at a new normal: where verification is paramount and software is patched continuously.]]></description><pubDate>Thu, 23 Apr 2026 14:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/ai-cybersecurity-mythos</guid><category>Cybersecurity</category><category>Anthropic</category><category>Agentic-ai</category><category>Hacking</category><dc:creator>Bruce Schneier</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-cgi-image-of-a-translucent-padlock-filled-with-0s-and-1s-one-spot-is-broken-and-the-numbers-are-spraying-out-of-that-spot.jpg?id=65714765&amp;width=980"></media:content></item><item><title>AI Agent Designs a RISC-V CPU Core From Scratch</title><link>https://spectrum.ieee.org/ai-chip-design</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-graphic-design-system-plot-of-a-risc-v-cpu-core-it-resembles-a-square-grid-covered-in-colorful-vertical-and-horizontal-scratc.jpg?id=65519361&width=1245&height=700&coordinates=0%2C469%2C0%2C469"/><br/><br/><p>In 2020, researchers fine-tuned a GPT-2 model to <a href="https://arxiv.org/html/2411.11856v2" rel="noopener noreferrer" target="_blank">design fragments of logic circuits</a>; in 2023, researchers used GPT-4 <a href="https://arxiv.org/abs/2305.13243" rel="noopener noreferrer" target="_blank">to help design an 8-bit processor</a> with a novel instruction set; by 2024, a variety of LLMs could <a href="https://arxiv.org/pdf/2405.02326" rel="noopener noreferrer" target="_blank">design and test chips</a> with basic functionality, like dice rolls (though often these were flawed).</p><p>Now Verkor.io, an <a href="https://spectrum.ieee.org/chip-design-ai" target="_blank">AI chip design</a> startup, claims a bigger milestone: a <a href="https://spectrum.ieee.org/risc-v-laptops" target="_blank">RISC-V </a>CPU core designed entirely by an agentic AI system. The CPU, dubbed VerCore, has a clock speed of 1.5 gigahertz and performance similar to a 2011-era laptop CPU. </p><p><a href="https://www.linkedin.com/in/suresh-krishna-793506158" rel="noopener noreferrer" target="_blank">Suresh Krishna</a>, cofounder at <a href="https://verkor.io/" rel="noopener noreferrer" target="_blank">Verkor.io</a>, says the team’s key claim is that this approach is more effective than using only specialized AI systems for specialized tasks within the overall design process. “ What we learned is that the better approach is to let the AI agent solve the whole problem,” he says.</p><h2>Bringing Human Workflows to Agentic AI</h2><p>Verkor.io’s agentic system is called <a href="https://arxiv.org/pdf/2603.08716" rel="noopener noreferrer" target="_blank">Design Conductor</a>, and it’s not itself an AI model. It’s a harness for large language models (LLMs). A harness is software that forces an AI agent to proceed through structured steps. In this case, the steps are like those a team of human chip architects would follow: design, implementation, testing, and so on. The harness also manages subagents and a database of related files.</p><p>That means it can work autonomously with only an initial prompt—in this case a 219-word design specification—from the user. (<a href="https://arxiv.org/pdf/2603.08716" target="_blank">The prompt is published in the Design Conductor paper</a>.) It outputs <a href="https://en.wikipedia.org/wiki/GDSII" rel="noopener noreferrer" target="_blank">a Graphic Design System II (GDSII) file</a>, which can be used in existing electronic design automation (EDA) software.</p><p><a href="https://www.synopsys.com/ai/agentic-ai.html" rel="noopener noreferrer" target="_blank">Synopsys</a> and <a href="https://www.cadence.com/en_US/home/ai/ai-for-design.html" rel="noopener noreferrer" target="_blank">Cadence</a>, two major players in EDA software, also have agentic AI tools. These allow chip architects to automate some tasks with AI agents. Design Conductor is different because it’s built to handle chip design from spec to completion with full autonomy, something major EDA companies have not yet touted.</p><p><a href="https://www.linkedin.com/in/ravi-k-a10287122/" target="_blank">Ravi Krishna</a>, founding engineer at Verkor.io, says Design Conductor’s workflow is “mirrored after the traditional process a human engineer might use.” It analyzes the specification, then writes and debugs a register-transfer level, or RTL, file (an abstraction of the CPU’s data flow) before iterating through subtasks like power delivery, signal timings, and layout, which are again checked against the specification. Some tasks, like layout, <a href="https://theopenroadproject.org/" target="_blank">call tools</a> to assist the agent. “It’s an iterative system.”</p><p>The system took 12 hours to create the VerCore design. That’s not long, but, because it uses AI agents, you might imagine it taking more or less time based on the number of agents thrown at it. However, Ravi Krishna says it’s not that simple, because some design tasks aren’t easily parallelized. </p><p>However, the general improvement of AI models over time has proven essential. “I remember that around the middle of last year, we tried to build a floating-point multiplier with the models of that time. It was slightly beyond what they could do,” says Ravi Krishna. VerCore—designed in December 2025— represents an increase in capability since then. “If it can’t do it today, it’ll do it in six months,” he says. “I don’t know if that’s a scary thing or a good thing.”</p><h2>A First for AI Chip Design</h2><p>VerCore uses the RISC-V instruction set architecture (ISA), a popular open-standard ISA that’s beginning to break out of niche applications, like storage controllers, into systems on a chip (SoCs) that can power <a href="https://spectrum.ieee.org/risc-v-laptops" target="_self">laptops or smartphones</a>. The CPU’s exact clock speed is 1.48 GHz and it achieved a <a href="https://www.eembc.org/coremark/" rel="noopener noreferrer" target="_blank"></a>score of 3,261 on the <a href="https://www.eembc.org/coremark/" rel="noopener noreferrer" target="_blank">CoreMark</a> processor core benchmark. </p><p>Verkor says this puts VerCore’s performance in line with the CPU core performance of <a href="https://www.notebookcheck.net/Intel-Celeron-Dual-Core-SU2300-Notebook-Processor.33847.0.html" rel="noopener noreferrer" target="_blank">Intel’s Celeron SU2300</a>. Whether that sounds impressive depends on your perspective. The Celeron SU2300, which arrived in 2011, uses Intel’s <a href="https://www.intel.com/content/dam/doc/white-paper/45nm-next-generation-core-microarchitecture-white-paper.pdf" rel="noopener noreferrer" target="_blank">Penryn CPU architecture</a>, which debuted in November of 2007.<br/><br/> In other words, VerCore is no threat to leading-edge CPUs, but it’s notable for two reasons.<br/><br/>VerCore is the first RISC-V CPU core designed by an AI agent. Previous examples of AI chip design presented portions of a design but didn’t present a complete core. Ravi Krishna says the company wanted to target a design that an AI agent hadn’t previously accomplished. “From the perspective of trying to push the limits of what AI models can do, that was interesting to us,” he says.</p><p>And while VerCore’s theoretical performance has limits, it’s enough to suggest the design could be useful. Indeed, RISC-V is popular because it provides an ISA that’s free to use (RISC-V is an open standard). RISC-V chips generally aren’t as quick as their <em>x</em>86 and Arm peers, but they’re less expensive. </p><p>There’s one final caveat worth mentioning; the chip has not been physically produced. VerCore was verified in simulation with <a href="https://github.com/riscv-software-src/riscv-isa-sim" rel="noopener noreferrer" target="_blank">Spike</a>, the reference RISC-V ISA simulator, and laid out using the open-source <a href="https://github.com/The-OpenROAD-Project/asap7" rel="noopener noreferrer" target="_blank">ASAP7 PDK</a>, an academic design kit that simulates a 7-nanometer production node. Both tools are commonly used for RISC-V design. VerCore says its CPU can run a variant of <a href="https://en.wikipedia.org/wiki/%CE%9CClinux" rel="noopener noreferrer" target="_blank">uCLinux</a> in simulation. </p><p>Skeptics will have a chance to judge for themselves. Verkor.io plans to release design files at the end of April. This will include the VerCore CPU and several other designs recently completed by the AI agent system. Verkor also plans to show an FPGA implementation of VerCore at <a href="https://dac.com/2026" rel="noopener noreferrer" target="_blank">DAC</a>, the leading electronic design automation conference.</p><h2>Should Chip Designers Worry about AI Agents Taking Their Jobs?</h2><p>An AI chip designer that can bang out a CPU in 12 hours might seem like troubling news for flesh-and-blood engineers, but Design Conductor has its limitations. The team at Verkor.io say that despite improvements, LLMs still lack the intuition a human can bring.</p><p>Design Conductor can fall down rabbit holes that a human engineer would avoid. In one instance the agent made a mistake in timing, meaning that data was not moved across the CPU in agreement with its clock cycle. The model didn’t recognize the cause and made broad changes while hunting for the fix. It did eventually find a fix, but only after reaching many dead ends. “Basically, we are trading off experience for compute,” says <a href="https://www.linkedin.com/in/david-chin-a5092a/" rel="noopener noreferrer" target="_blank">David Chin</a>, vice president of engineering at the startup.<br/><br/>Suresh Krishna concurs and adds that Design Conductor’s brute-force approach is likely to become less efficient as agentic systems tackle more complex designs. “It’s a nonlinear design space, so the compute grows very quickly,” he says. “As a practical matter, expert guidance and common sense helps a lot.”</p><p>Despite such issues, agentic systems like Design Conductor might accelerate chip design by accelerating iteration. They may also make design accessible to small teams that otherwise lack the resources or head count to pull off a project.</p><p>“It’s not at the point where you can have one person. I would say you still need five to ten, all experts in different areas,” says Ravi Krishna. “That team could get you to [a production-ready chip design] at this point.”</p>]]></description><pubDate>Wed, 22 Apr 2026 11:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/ai-chip-design</guid><category>Eda</category><category>Chip-design</category><category>Agentic-ai</category><category>Risc-v</category><category>Cpu</category><dc:creator>Matthew S. Smith</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-graphic-design-system-plot-of-a-risc-v-cpu-core-it-resembles-a-square-grid-covered-in-colorful-vertical-and-horizontal-scratc.jpg?id=65519361&amp;width=980"></media:content></item><item><title>Designing Broadband LPDA-Fed Reflector Antennas With Full-Wave EM Simulation</title><link>https://content.knowledgehub.wiley.com/efficient-design-and-simulation-of-lpda-fed-parabolic-reflector-antennas/</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/wipl-d-logo.png?id=26851496&width=980"/><br/><br/><p>A practical guide to designing log-periodic dipole array fed parabolic reflector antennas using advanced 3D MoM simulation — from parametric modeling to electrically large structures.</p><p><strong>What Attendees will Learn</strong></p><ol><li>How to set design requirements for LPDA-fed reflector antennas — Understand the key specifications including bandwidth ratio, gain targets, and VSWR matching constraints across the full operating range from 100 MHz to 1 GHz.</li><li>Why advanced 3D EM solvers enable simulation of electrically large multiscale structures — Learn how higher order basis functions, quadrilateral meshing, geometrical symmetry, and CPU/GPU parallelization extend MoM simulation capability by an order of magnitude.</li><li>How to apply a systematic three-step design strategy with proven workflow starting with first optimizing the stand-alone LPDA for VSWR and gain, then integrating the reflector, and finally tuning parameters to satisfy all performance requests including gain and impedance matching.</li><li>How parametric CAD modeling accelerates LPDA design — Discover how self-scaling geometry, automated wire-to-solid conversion, and multiple-copy-with-scaling features enable fully parametrized antenna models that streamline optimization across dozens of design variants.</li></ol><div><span><a href="https://content.knowledgehub.wiley.com/efficient-design-and-simulation-of-lpda-fed-parabolic-reflector-antennas/" target="_blank">Download this free whitepaper now!</a></span></div>]]></description><pubDate>Fri, 17 Apr 2026 14:00:50 +0000</pubDate><guid>https://content.knowledgehub.wiley.com/efficient-design-and-simulation-of-lpda-fed-parabolic-reflector-antennas/</guid><category>Type-whitepaper</category><category>Broadband</category><category>Antennas</category><category>Simulation</category><dc:creator>WIPL-D</dc:creator><media:content medium="image" type="image/png" url="https://assets.rbl.ms/26851496/origin.png"></media:content></item><item><title>Stealth Signals Are Bypassing Iran’s Internet Blackout</title><link>https://spectrum.ieee.org/iran-internet-blackout-satellite-tv</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/image.png?id=65716479&width=1245&height=700&coordinates=0%2C700%2C0%2C701"/><br/><br/><p><strong>On 8 January 2026, </strong>the Iranian government imposed a near-total communications shutdown. It was the country’s first full information blackout: For weeks, the internet was off across all provinces while services including the government-run intranet, VPNs, text messaging, mobile calls, and even landlines were severely throttled. It was an unprecedented lockdown that left more than <a href="https://www.chathamhouse.org/2026/01/irans-internet-shutdown-signals-new-stage-digital-isolation" rel="noopener noreferrer" target="_blank">90 million people</a> cut off not only from the world, but from one another.</p><div class="rm-embed embed-media"><iframe height="110px" id="noa-web-audio-player" src="https://embed-player.newsoveraudio.com/v4?key=q5m19e&id=https://spectrum.ieee.org/iran-internet-blackout-satellite-tv&bgColor=F5F5F5&color=1b1b1c&playColor=1b1b1c&progressBgColor=F5F5F5&progressBorderColor=bdbbbb&titleColor=1b1b1c&timeColor=1b1b1c&speedColor=1b1b1c&noaLinkColor=556B7D&noaLinkHighlightColor=FF4B00&feedbackButton=true" style="border: none" width="100%"></iframe></div><p>Since then, connectivity has never fully returned. Following <a href="https://en.wikipedia.org/wiki/2026_Iran_war" rel="noopener noreferrer" target="_blank">U.S. and Israeli airstrikes</a> in late February, Iran again imposed near-total restrictions, and people inside the country again saw global information flows dry up.</p><p>The original January shutdown came amid nationwide protests over the deepening economic crisis and political repression, in which millions of people chanted antigovernment slogans in the streets. While Iranian protests have become frequent in recent years, this was one of the most significant uprisings since the Islamic Revolution in 1979. The government responded quickly and brutally. One report put the death toll at <a href="https://www.en-hrana.org/the-crimson-winter-a-50-day-record-of-irans-2025-2026-nationwide-protests/" rel="noopener noreferrer" target="_blank">more than 7,000 confirmed deaths</a> and more than 11,000 under investigation. Many sources believe the death toll could exceed 30,000.</p><p>Thirteen days into the January shutdown, we at <a href="https://www.netfreedompioneers.org/" rel="noopener noreferrer" target="_blank">NetFreedom Pioneers</a> (NFP) turned to a system we had built for exactly this kind of moment—one that sends files over ordinary satellite TV signals. During the national information vacuum, our technology, called <a href="https://www.netfreedompioneers.org/toosheh-datacasting-technology/" rel="noopener noreferrer" target="_blank">Toosheh</a>, delivered real-time updates into Iran, offering a lifeline to millions starved of trusted information.</p><h2>How Iran Censors the Internet<br/></h2><p>I joined NetFreedom Pioneers, a nonprofit focused on anticensorship technology, in 2014. Censorship in <a href="https://spectrum.ieee.org/tag/iran" target="_blank">Iran</a> was a defining feature of my youth in the 1990s. After the Islamic Revolution, most Iranians began to lead double lives—one at home, where they could drink, dance, and choose their clothing, and another in public, where everyone had to comply with stifling government laws.</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="Photo of a helmeted soldier with a machine gun standing in front of an Iranian flag and cell tower." class="rm-shortcode" data-rm-shortcode-id="ef533f84cc5eb097a4cfe78e30b2984b" data-rm-shortcode-name="rebelmouse-image" id="7a368" loading="lazy" src="https://spectrum.ieee.org/media-library/photo-of-a-helmeted-soldier-with-a-machine-gun-standing-in-front-of-an-iranian-flag-and-cell-tower.jpg?id=65520617&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">Iran’s internet infrastructure is more centralized than in other parts of the world, making it easier for the government to restrict the flow of information. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Morteza Nikoubazl/NurPhoto/Getty Images</small></p>My first experience with secret communications was when I was five and living in the small city of Fasa in southern Iran. My uncle brought home a satellite dish—dangerously illegal at the time—that allowed us to tune into 12 satellite channels. My favorite was Cartoon Network. Then, during my teenage years, this same uncle introduced me to the internet through dial-up modems. I remember using Yahoo Mail with its 4 megabytes of storage, reading news from around the world, and learning about the Chandra X-ray telescope from NASA’s website. <p><br/><br/><span>That openness didn’t last. As internet use spread in the early 2000s, the Iranian government began reshaping the network itself. Unlike the highly distributed networks in the United States or Europe, where thousands of providers exchange traffic across many independent routes, Iran’s connection to the global internet is relatively centralized. Most international traffic passes through a small number of gateways controlled by state-linked telecom operators. That architecture gives authorities unusual leverage: By restricting or withdrawing those connections, they can sharply reduce the country’s access to the outside world.</span></p><p>Over the past decade, Iran has expanded this control through what it calls the <a href="https://en.wikipedia.org/wiki/National_Information_Network" target="_blank">National Information Network</a>, a domestically routed system designed to keep data inside the country whenever possible. Many government services, banking systems, and local platforms are hosted on this internal network. During periods of unrest, access to the global internet can be throttled or cut off while portions of this domestic network continue to function.</p><h3></h3><br/><div class="rblad-ieee_in_content"></div><p>The government began its censorship campaign by redirecting or blocking websites. As internet use grew, it adopted more sophisticated approaches. For example, the <a href="https://en.wikipedia.org/wiki/Telecommunication_Company_of_Iran" target="_blank">Telecommunication Company of Iran</a> uses a technique called <a href="https://www.fortinet.com/resources/cyberglossary/dpi-deep-packet-inspection" target="_blank">deep packet inspection</a> to analyze the content of data packets in real time. This method enables it to identify and block specific types of traffic, such as VPN connections, messaging apps, social media platforms, and banned websites.</p><h2>The Stealth of Satellite Transmissions<br/></h2><p>Toosheh’s communication workaround builds on a history of satellite TV adoption in Middle Eastern and North African countries. By the early 2000s, satellite dishes were common in Iran; today the majority of households in Iran have access to satellite TV despite its official prohibition.</p><p>Unlike subscription services such as DirecTV and Dish Network, “free-to-air” satellite TV broadcasts are unencrypted and can be received by anyone with a dish and receiver—no subscription required. Because the signals are open, users can also capture and store the data they carry, rather than simply watching it live. Tech-savvy people learned that they could use a digital video broadcasting (DVB) card—a piece of hardware that connects to a computer and tunes into satellite frequencies—to transform a personal computer into a satellite receiver. This way, they could watch and store media locally as well as download data from dedicated channels.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Photo of satellite dishes adorning the side of an apartment building." class="rm-shortcode" data-rm-shortcode-id="a558326e8ca2bd5c645e392fb0166b58" data-rm-shortcode-name="rebelmouse-image" id="577d2" loading="lazy" src="https://spectrum.ieee.org/media-library/photo-of-satellite-dishes-adorning-the-side-of-an-apartment-building.jpg?id=65520620&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">Many Iranian citizens have free-to-air satellite dishes, like the ones on this apartment building in Tehran, and can thus download Toosheh transmissions, giving them a lifeline during internet blackouts.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Morteza Nikoubazl/NurPhoto/Getty Images</small></p><p>Toosheh, a Persian word that translates to “knapsack,” is the brainchild of <a href="https://x.com/mehdiy_fa" target="_blank">Mehdi Yahyanejad</a>, an Iranian-American technologist and entrepreneur. Yahyanejad cofounded NetFreedom Pioneers in 2012. He proposed that the satellite-computer connections enabled by a DVB card could be re-created in software, eliminating the need for specialized hardware. He added a simple digital interface to the software to make it easy for anyone to use. The next breakthrough came when the NFP team developed a new transfer protocol that tricks ordinary satellite receivers into downloading data alongside audio and video content. Thus, Toosheh was born.</p><p>Satellite TV uses a file system called an <a href="https://en.wikipedia.org/wiki/MPEG_transport_stream" target="_blank">MPEG transport stream</a> that allows multiple audio, video, or data layers to be packaged into a single stream file. When you tune in to a satellite channel and select an audio option or closed captions, you’re accessing data stored in different parts of this stream. The NFP team’s insight was that, by piggybacking on one of these layers, Toosheh could send an MPEG stream that included documents, videos, and more.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="An illustration of an 8 step process for sending digital files via satellite TV signals." class="rm-shortcode" data-rm-shortcode-id="500fc02c0c38f890606e42dec590ae8f" data-rm-shortcode-name="rebelmouse-image" id="371ea" loading="lazy" src="https://spectrum.ieee.org/media-library/an-illustration-of-an-8-step-process-for-sending-digital-files-via-satellite-tv-signals.png?id=65521138&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">HOW TOOSHEH WORKS: At NetFreedom Pioneers, content curators pull together files—news articles, videos, audio, and software [1]. Toosheh’s encoder software [2] compresses the files into a bundle, in .ts format, creating an MPEG transport stream [3]. From there, it’s uploaded to a server for transmission [4] via a free-to-air TV channel on a Yahsat satellite that’s positioned over the Middle East to provide regional coverage [5]. Satellite receivers [6] directly capture the data streams, which are downloaded to computers, smartphones, and other devices, and decoded by Toosheh software [8].</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Chris Philpot</small></p><p>A satellite receiver can’t tell the difference between our data and normal satellite audio and video data since it only “sees” the MPEG streams, not what’s encoded on them. This means the data can be downloaded and read, watched, and saved on local devices such as computers, smartphones, or storage devices. What’s more, the system is entirely private: No one can detect whether someone has received data through Toosheh; there are no traceable logs of user activity.</p><p>Toosheh doesn’t provide internet access, but rather delivers curated data through satellite technology. The fundamental distinction lies in the way users interact with the system. Unlike traditional internet services, where you type a request into your browser and receive data in response, Toosheh operates more like a combination of radio and television, presenting information in a magazine-like format. Users don’t make requests; instead, they receive 1 to 5 gigabytes of prepackaged, carefully selected data.</p><p class="pull-quote"><span>Access to information is not only about news or politics, but about exposure to possibilities.  </span></p><p>During this year’s internet blackout, we distributed official statements from Iranian opposition leader Crown Prince Reza Pahlavi and the U.S. government. We provided first-aid tutorials for medics and injured protesters. We sent uncensored news reports from BBC Persian, Iran International, IranWire, VOA Farsi, and others. We also shared critical software packages including anticensorship and antisurveillance tools, along with how-to guides to help people securely connect to Starlink satellite terminals, allowing them to stay protected and anonymous as they sent their own communications.</p><h2>How to Combat Signal Interference<br/></h2><p>Because Toosheh relies on one-way satellite broadcasts, it evades the usual tactics governments use to block internet access. However, it remains vulnerable to <a href="https://spectrum.ieee.org/satellite-jamming" target="_blank">satellite signal jamming</a>.</p><p>The Iranian government is notorious for deploying signal jamming, especially in larger cities. In 2009, the government <a href="https://www.dw.com/fa-ir/%D9%86%D8%A7%D8%AA%D9%88%D8%A7%D9%86%DB%8C-%D8%AF%D8%B1-%D9%85%D9%82%D8%A7%D8%A8%D9%84-%D8%A7%D9%85%D9%88%D8%A7%D8%AC-%D9%BE%D8%A7%D8%B1%D8%A7%D8%B2%DB%8C%D8%AA-%D8%A7%D8%B2-%D8%AA%D9%87%D8%B1%D8%A7%D9%86/a-5417209" target="_blank">used uplink interference</a>, which attacks the satellite in orbit by beaming strong noise in the frequency of the satellite’s receiver. This makes it impossible for the satellite to distinguish the information it’s supposed to receive. However, because this type of attack temporarily disables the entire satellite, Iran was threatened with international <a href="https://www.dw.com/fa-ir/%D8%AA%D8%B4%D8%AF%DB%8C%D8%AF-%D8%A7%D9%86%D8%AA%D9%82%D8%A7%D8%AF%D9%87%D8%A7-%D8%A8%D9%87-%D8%A7%D8%B1%D8%B3%D8%A7%D9%84-%D9%BE%D8%A7%D8%B1%D8%A7%D8%B2%DB%8C%D8%AA-%D8%A7%D8%B2-%D8%B3%D9%88%DB%8C-%D8%A7%DB%8C%D8%B1%D8%A7%D9%86/a-5382663" target="_blank">sanctions</a> and in 2012 stopped using the method .</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A chart displayed on a cellphone shows internet connectivity in Iran dropped from almost 100% to 0% on 9 January 2026." class="rm-shortcode" data-rm-shortcode-id="c5f3ef2e60cfa653b7c461cda6d68e0f" data-rm-shortcode-name="rebelmouse-image" id="c778a" loading="lazy" src="https://spectrum.ieee.org/media-library/a-chart-displayed-on-a-cellphone-shows-internet-connectivity-in-iran-dropped-from-almost-100-to-0-on-9-january-2026.jpg?id=65520652&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">A graph of network connectivity in Iran shows that on 9 January 2026, internet access dropped from nearly 100 percent to 0. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Samuel Boivin/NurPhoto/Getty Images</small></p><p>The current method, called terrestrial jamming, uses antennas installed at higher elevations than the surrounding buildings to beam strong noise over a specific area in the frequency range of household receivers. This attack is effective in keeping some of the packets from arriving and damaging others, effectively jamming the transmission. But it’s short-range and requires significant power, so it’s impossible to implement nationwide. There are always people somewhere who can still watch TV, download from Toosheh, or tune into a satellite radio despite the jamming. Even so, we wanted a workaround that would keep our transmissions broadly accessible.</p><p>NFP’s solution was to add redundancy, similar in principle to a data-storage technique called RAID (redundant array of independent disks). Instead of sending each piece of data once, we send extra information that allows missing or corrupted packets to be reconstructed. Under normal circumstances, we often use 5 percent of our bandwidth for this redundancy. During periods of active jamming, we increase that to as much as 25 to 30 percent, improving the chances that users can recover complete files despite interference.</p><h2>From Crisis Response to Public Access<br/></h2><p>Toosheh initially came online in 2015 in Iran and Afghanistan. Its full potential, however, was first realized during the 2019 protests in Iran, which saw the most widespread internet shutdown prior to the blackout this year. <a href="https://www.wired.com/story/iran-news-internet-shutdown/" target="_blank"><em><em>Wired</em></em></a> called the 2019 shutdown “the most severe disconnection” tracked by <a href="https://netblocks.org/" target="_blank">NetBlocks</a> in any country in terms of its “technical complexity and breadth.” Our technology helped thousands of people stay informed. We sent crucial local updates, legal-aid guides, digital security tools, and independent news to satellite receivers all over the country, seeing a sixfold increase in our user base.</p><p>When that wave of protests subsided, the government allowed some communication services to return. People were again able to access the free internet using VPNs and other antifilter software that allowed them to bypass restrictions. Toosheh then became a public access point for news, educational material, and entertainment beyond government filtering.</p><p>Toosheh’s impact is often personal. A traveling teacher in western Iran told NFP that he regularly distributed Toosheh files to students in remote villages. One package included footage of female athletes competing in the Olympic Games, something never broadcast in Iran. For one young girl, it was the first time she realized women could compete professionally in sports. That moment underscores a broader truth: Access to information is not only about news or politics, but about exposure to possibilities.</p><h2>The Cost of Toosheh<br/></h2><p>Unlike internet-based systems, Toosheh’s operational cost remains constant regardless of the number of users. A single TV satellite in geostationary earth orbit, deployed and maintained by an international company such as Eutelsat, can broadcast to an entire continent with no increase in cost to audiences. What’s more, the startup cost for users isn’t high: A satellite dish and receiver in Iran costs less than US $50, which is affordable to many. And it costs nothing for people to use Toosheh’s service and receive its files.</p><p class="pull-quote"><span>We aim not just to build a tool for censorship circumvention, but to redefine access itself. </span></p><p>However, operating the service is costly: NetFreedom Pioneers pays tens of thousands of dollars a month for satellite bandwidth. We had received funding from the U.S. State Department, but in August of 2025, that funding ended, forcing us to suspend services in Iran.</p><p>Then the December protests happened, and broadcasting to Iran became an urgent priority. To turn Toosheh back on, we needed roughly $50,000 a month. With the support of a handful of private donors, we were able to meet these costs and sustain operations in Iran for a few months, though our future there and elsewhere is uncertain.</p><h2>Satellites Against Censorship<br/></h2><p>Toosheh’s revival in Iran came alongside NFP’s ongoing support for deployments of Starlink, a satellite internet service that allows users to connect directly to satellites rather than relying on domestic networks, which the government can shut down. Unlike Toosheh’s one-way broadcasts, <a href="https://spectrum.ieee.org/tag/starlink" target="_blank">Starlink</a> provides full two-way internet access, enabling users to send messages, upload videos, and communicate with the outside world.</p><p>In 2022, we started gathering <a href="https://www.gofundme.com/f/urgent-help-deliver-starlink-and-vpn-access-for-freedom" target="_blank">donations</a> to buy Starlink terminals for Iran. We have delivered more than 300 of the <a href="https://www.theguardian.com/world/2026/jan/13/ecosystem-smuggled-tech-iran-last-link-outside-world-internet" target="_blank">roughly 50,000</a> there, enabling citizens to send encrypted updates and videos to us from inside the country. Because the technology is banned by the government, access remains limited and carries risk; Iranian authorities have recently arrested Starlink users and sellers. And unlike Toosheh’s receive-only broadcasts, Starlink terminals transmit signals back to orbit, creating a radio footprint that can potentially be detected.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A photo of a laptop screen says the user is offline." class="rm-shortcode" data-rm-shortcode-id="2c0caa05d5589d7d25beeb8342db442e" data-rm-shortcode-name="rebelmouse-image" id="103c7" loading="lazy" src="https://spectrum.ieee.org/media-library/a-photo-of-a-laptop-screen-says-the-user-is-offline.png?id=65521782&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">The internet shutdown in Iran continued after the attacks by Israel and the United States began in late February, preventing Iranians from communicating with the outside world and with one another.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Fatemeh Bahrami/Anadolu/Getty Images</small></p><p>Looking ahead, we envision Toosheh becoming a foundational part of global digital resilience. It is uncensored, untraceable, and resistant to government shutdowns. Because Toosheh is downlink only, it can sometimes feel hard to explain the value of this technology to those living in the free world, those accustomed to open internet access. Yet, people living under censorship have few other choices when there’s a digital blackout.</p><p>Currently, NFP is developing new features like intelligent content curation and automatically prioritizing data packages based on geographic or situational needs. And we’re experimenting with local sharing tools that allow users who receive Toosheh broadcasts to redistribute those files via Wi-Fi hotspots or other offline networks, which could extend the system’s reach to disaster zones, conflict areas, and climate-impacted regions where infrastructure may be destroyed.</p><p>We’re also looking at other use cases. Following the Taliban’s return to power in Afghanistan, NetFreedom Pioneers designed a satellite-based system to deliver educational materials. Our goal is to enable private, large-scale distribution of coursework to anyone—including the girls who are banned from Afghanistan’s schools. The system is technically ready but has yet to secure funding for deployment.</p><p>We aim not just to build a tool for censorship circumvention, but to redefine access itself. Whether in an Iranian city under surveillance, a Guatemalan village without internet, or a refugee camp in East Africa, Toosheh offers a powerful and practical model for delivering vital information without relying on vulnerable or expensive networks.</p><p>Toosheh is a reminder that innovation doesn’t have to mean complexity. Sometimes, the most transformative ideas are the simplest, like delivering data through the sky, quietly and affordably, into the hands of those who need it most.<span class="ieee-end-mark"></span></p><p><em>This article appears in the May 2026 print issue as “The Stealth Signals Bypassing Iran’s Internet Blackout.”</em></p>]]></description><pubDate>Wed, 15 Apr 2026 13:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/iran-internet-blackout-satellite-tv</guid><category>Satellite-communications</category><category>Censorship</category><category>Iran</category><category>Protests</category><category>Democracy</category><category>Internet-shutdowns</category><dc:creator>Evan Alireza Firoozi</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/image.png?id=65716479&amp;width=980"></media:content></item><item><title>Crypto Faces Increased Threat From Quantum Attacks</title><link>https://spectrum.ieee.org/quantum-safe-crypto</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/abstract-pixel-art-resembling-a-padlock-and-token.jpg?id=65520763&width=1245&height=700&coordinates=0%2C156%2C0%2C157"/><br/><br/><p>The <a href="https://spectrum.ieee.org/post-quantum-cryptography-standards-nist" target="_self">race</a> to transition online security protocols to ones that can’t be cracked by a quantum computer is already on. The algorithms that are commonly used today to protect data online—<a href="https://en.wikipedia.org/wiki/RSA_cryptosystem" rel="noopener noreferrer" target="_blank">RSA</a> and <a href="https://en.wikipedia.org/wiki/Elliptic-curve_cryptography" rel="noopener noreferrer" target="_blank">elliptic curve cryptography</a>—are uncrackable by supercomputers, but a large enough quantum computer would make quick work of them. There are <a href="https://spectrum.ieee.org/post-quantum-cryptography-2668949802" target="_self">algorithms</a> secure enough to be out of reach for both classical and future quantum machines, called post-quantum cryptography, but transitioning to these is a <a href="https://spectrum.ieee.org/post-quantum-cryptography-2667758178" target="_self">work in progress</a>. </p><p>Late last month, the team at <a href="https://quantumai.google/" rel="noopener noreferrer" target="_blank">Google Quantum AI</a> published a <a href="https://arxiv.org/abs/2603.28846" rel="noopener noreferrer" target="_blank">whitepaper</a> that added significant urgency to this race. In it, the team showed that the size of a quantum computer that would pose a cryptographic threat is approximately 20 times <a href="https://research.google/blog/safeguarding-cryptocurrency-by-disclosing-quantum-vulnerabilities-responsibly/" rel="noopener noreferrer" target="_blank">smaller</a> than previously thought. This is still far from accessible to the quantum computers that exist today: The largest machines currently consist of approximately 1,000 quantum bits, or qubits, and the whitepaper estimated that about 500 times as much is needed. Nonetheless, this shortens the timeline to switch over to post-quantum algorithms. </p><p>The news had a surprising beneficiary: Obscure cryptocurrency <a href="https://algorand.co/" rel="noopener noreferrer" target="_blank">Algorand</a> <a href="https://www.indexbox.io/blog/algorand-price-surges-44-after-google-research-paper-citation/" rel="noopener noreferrer" target="_blank">jumped</a> 44% in price in response. The whitepaper called out Algorand specifically for implementing post-quantum cryptography on their blockchain. We caught up with Algorand’s chief scientific officer and professor of computer science and engineering at the University of Michigan, <a href="https://web.eecs.umich.edu/~cpeikert/" rel="noopener noreferrer" target="_blank">Chris Peikert</a>, to understand how this announcement is impacting cryptography, why cryptocurrencies are feeling the effects, and what the future might hold. Peikert’s early work on a particular type of algorithm known as <a href="https://en.wikipedia.org/wiki/Lattice-based_cryptography" rel="noopener noreferrer" target="_blank">lattice cryptography</a> underlies most post-quantum security today.</p><p><strong>IEEE Spectrum:</strong><span> What is the significance of this Google Quantum AI whitepaper?</span></p><p><strong>Peikert:</strong> The upshot of this paper is that it shows that a quantum computer would be able to break some of the cryptography that is most widely used, especially in blockchains and cryptocurrencies, with much, much fewer resources than had previously been established. Those resources include the time that it would take to do so and the number of qubits (or quantum bits) that it would have to use.</p><p>This cryptography is very central to not just cryptocurrencies, but more broadly to cryptography on the internet. It is also used for secure web connections between web browsers and web servers. Versions of elliptic curve cryptography are used in national security systems and military encryption. It’s very prevalent and pervasive in all modern networks and protocols.</p><p>And not only was this paper improving the algorithms, but there was also a concurrent paper showing that the hardware itself was substantially improved. The claim here was that the number of physical qubits needed to achieve a certain kind of logical qubit was also greatly reduced. These two kinds of improvements are compounding upon each other. It’s a kind of a win-win situation from the quantum computing perspective, but a lose-lose situation for cryptography.</p><p><strong>IEEE Spectrum: </strong>What do Google AI’s findings mean for cryptocurrencies and the broader cybersecurity ecosystem?</p><p><strong>Peikert:</strong> There’s always been this looming threat in the distance of quantum computers breaking a large fraction of the cryptography that’s used throughout the cryptocurrency ecosystem. And I think what this paper did was really the loudest alarm yet that these kinds of quantum attacks might not be as far off as some have suspected, or hoped, in recent years. It’s caused a reevaluation across the industry, and a moving up of the timeline for when quantum computers might be capable of breaking this cryptography.</p><p>When we think about the timelines and when it’s important to have completed these transitions [to post-quantum cryptography], we also need to factor in the unknown improvements that we should expect to see in the coming years. The science of quantum computing will not stay static, and there will be these further breakthroughs. We can’t say exactly what they will be or when they will come, but you can bet that they will be coming.</p><p><strong>IEEE Spectrum:</strong> What is your guess on if or when quantum computers will be able to break cryptography in the real world?</p><p><strong>Peikert:</strong> Instead of thinking about a specific date when we expect them to come, we have to think about the probabilities and the risks as time goes on. There have been huge breakthrough developments, including not only this paper, but also <a href="https://research.google/blog/making-quantum-error-correction-work/" target="_blank">some</a> last year. But even with these, I think that the chance of a cryptographic attack by quantum computers being successful in the next three years is extremely low, maybe less than a percent. But then, as you get out to several years, like five, six, or 10 years, one has to seriously consider a probability, maybe 5 percent or 10 percent or more. So it’s still rather small, but significant enough that we have to worry about the risk, because the value that is protected by this kind of cryptography is really enormous. </p><p>The U.S. government has put 2035 as its target for migrating all of the national security systems to post-quantum cryptography. That seems like a prudent date, given the timelines that it takes to upgrade cryptography. It’s a slow process. It has to be done very deliberately and carefully to make sure that you’re not introducing new vulnerabilities, that you’re not making mistakes, that everything still works properly. So, you know, given the outlook for quantum computers on the horizon, it’s really important that we prepare now, or ideally, yesterday, or a few years ago, for that kind of transition.</p><p><strong>IEEE Spectrum: </strong>Are there significant roadblocks you see to industrial adoption of post-quantum cryptography going forward?</p><p><strong>Peikert:</strong> Cryptography is very hard to change. We’ve only had one or maybe two major transitions in cryptography since the early 1980s or late 1970s, when the field first was invented. We don’t really have a systematic way of transitioning cryptography. </p><p>An additional challenge is that the performance trade-offs are very different in post-quantum cryptography than they are in the legacy systems. Keys and cipher texts and digital signatures are all significantly larger in post-quantum cryptography, but the computations are actually faster, typically. People have optimized cryptography for speed in the past, and we have very good fast speeds now for post-quantum cryptography, but the sizes of the keys are a challenge. </p><p>Especially in blockchain applications, like cryptocurrencies, space on the blockchain is at a premium. So it calls for a reevaluation in many applications of how we integrate the cryptography into the system, and that work is ongoing. And, the blockchain ecosystem uses a lot of advanced cryptography, exotic things like zero-knowledge proofs. In many cases, we have rudimentary constructions of these fancy cryptography tools from post-quantum-type mathematics, but they’re not nearly as mature and industry-ready as the legacy systems that have been deployed. It continues to be an important technical challenge to develop post-quantum versions of these very fancy cryptographic schemes that are used in cutting-edge applications.</p><p><strong>IEEE Spectrum: </strong>As an academic cryptography researcher, what attracted you to work with a cryptocurrency, and Algorand in particular?</p><p><strong>Peikert:</strong> My former Ph.D. advisor is <a href="https://en.wikipedia.org/wiki/Silvio_Micali" target="_blank">Silvio Micali</a>, the inventor of Algorand. The system is very elegant. It is a very high-performing blockchain system, and it uses very little energy, has fast transaction finalization, and a number of other great features. And Silvio appreciated that this quantum threat was real and was coming, and the team approached me about helping to improve the Algorand protocol at the basic levels to become more post-quantum secure in 2021. That was a very exciting opportunity, because it was a difficult engineering and scientific challenge to integrate post-quantum cryptography into all the different technical and cryptographic mechanisms that were underlying the protocol.</p><p><strong>IEEE Spectrum: </strong>What is the current status of post-quantum cryptography in Algorand, and blockchains in general? </p><p><strong>Peikert:</strong> We’ve identified some of the most pressing issues and worked our way through some of them, but it’s a many-faceted problem overall. We started with the integrity of the chain itself, which is the transaction history that everybody has to agree upon. </p><p>Our first major project was developing a system that would add post-quantum security to the history of the chain. We developed a system called <a href="https://dev.algorand.co/concepts/protocol/state-proofs/" rel="noopener noreferrer" target="_blank">state proofs</a> for that, which is a mixture of ordinary post-quantum cryptography and also some more fancy cryptography: It’s a way of taking a large number of signatures and digesting them down into a much smaller number of signatures, while still being confident that these large number of signatures actually exist and are properly formed. We also followed it with other papers and projects that are about adding post-quantum cryptography and security to other aspects of the blockchain in the Algorand ecosystem. </p><p>It’s not a complete project yet. We don’t claim to be fully post-quantum secure. That’s a very challenging target to hit, and there are aspects that we will continue to work on into the near future.</p><p><strong>IEEE Spectrum: </strong>In your view, will we adopt post-quantum cryptography before the risks actually catch up with us? </p><p><strong>Peikert:</strong> I tend to be an optimist about these things. I think that it’s a very good thing that more people in decision-making roles are recognizing that this is an important topic, and that these kinds of migrations have to be done. I think that we can’t be complacent about it, and we can’t kick the can down the road much longer. But I do see that the focus is being put on this important problem, so I’m optimistic that most important systems will eventually have good either mitigations or full migrations in place. </p><p>But it’s also a point on the horizon that we don’t know exactly when it will come. So, there is the possibility that there is a huge breakthrough, and we have many fewer years than we might have hoped for, and that we don’t get all the systems upgraded that we would like to have fixed by the time quantum computers arrive.</p>]]></description><pubDate>Wed, 15 Apr 2026 13:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/quantum-safe-crypto</guid><category>Quantum-computing</category><category>Post-quantum-cryptography</category><category>Cryptocurrency</category><category>Lattice-cryptography</category><category>Security-protocols</category><category>Blockchain</category><category>Cryptography</category><dc:creator>Dina Genkina</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/abstract-pixel-art-resembling-a-padlock-and-token.jpg?id=65520763&amp;width=980"></media:content></item><item><title>Squishy Photonic Switches Promise Fast Low-Power Logic</title><link>https://spectrum.ieee.org/soft-photonics</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/illustration-of-a-micropipette-piercing-through-a-hemisphere-shaped-membrane-to-inject-a-droplet-at-its-core.jpg?id=65506297&width=1245&height=700&coordinates=0%2C187%2C0%2C188"/><br/><br/><p><span>Photonic devices, which rely on light instead of electricity, have the potential to be faster and more energy efficient than today’s electronics. They also present a unique opportunity to develop devices using <a href="https://spectrum.ieee.org/soft-robot-actuators-bugs" target="_self">soft materials</a>, such as polymers and gels, which are poor conductors of electricity but are easier to manufacture and more environmentally friendly. The development of these potentially squishy, <a href="https://spectrum.ieee.org/wearable-sensors" target="_self">flexible photonics</a>, however, requires the ability to manipulate light using only light, not electricity.</span></p><p>In soft matter, that’s been done primarily by changing the physical properties of optical materials or by using intense light pulses to change the direction of light. Now, an international team of scientists has developed a new way of controlling light with light using very low light intensities and without changing any of the physical properties of materials. </p><p><a href="https://musevic.fmf.uni-lj.si/" target="_blank"><span>Igor Muševič</span></a>, a professor of physics at the University of Ljubljana who led the project, says that he first got the idea for the device while at a conference in San Francisco, listening to a talk by <a href="https://www.nobelprize.org/prizes/chemistry/2014/hell/facts/" target="_blank">Stefan W. Hell </a>about stimulated emission depletion (STED) microscopy. The imaging technique, for which Hell won a <a href="https://www.nobelprize.org/prizes/chemistry/2014/summary/" target="_blank">Nobel Prize in Chemistry in 2014</a>, uses two lasers to produce an extremely small light beam to scan objects. “When I saw this, I said, This is manipulation light by light, right?” Muševič recalls.</p><p><span>His realization inspired a device into which a laser pulse is fired. Whether or not this beam makes it out of the device depends on whether or not a second pulse is fired less than a nanosecond afterwards.</span></p><h2>A liquid crystal photonic switch</h2><p><span>The device consists of a spherically shaped bead of liquid crystal, held in shape by its elastic material properties and the forces between its molecules, infused with a fluorescent dye and trapped between four upright cone-shaped polymer structures that guide light in and out of the device. When a laser pulse is sent through one of the four polymer waveguides, the light is quickly transferred into the liquid crystal, exciting the fluorescent dye. In a process known as whispering gallery mode resonance, the photons inside the liquid crystal are reflected back inside each time they hit the liquid’s spherical surface. The result is that light circulates inside the cavity until it is eventually reflected into one of the waveguides, which then emits the photons out in a laser beam. </span></p><p>The team realized that sending a second laser pulse of a different color into the waveguides before the liquid crystal started emitting light from the first laser pulse resulted in stimulated emission of the excited dye molecules. The photons from the second laser pulse, which had to be fired into the waveguides after the first laser pulse, interact with the already-excited dye molecules. The interaction causes the dye to emit photons identical to those in the second pulse while depleting the energy from the first pulse. The second laser beam, called the STED beam, is amplified by the process, while the light from the first pulse is so diminished that it isn’t emitted at all. Because the outcome of the first laser pulse could be controlled using the second laser pulse, the team had successfully demonstrated the control of light by light.</p><p class="shortcode-media shortcode-media-youtube"> <span class="rm-shortcode" data-rm-shortcode-id="0cb7a5df3d8c2896d2f429edfd746f29" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/mImgOT2zJ0I?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span> <small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Vandna Sharma, Jaka Zaplotnik, et al.</small> </p><p><span>According to the Ljubljana team, the energy efficiency of the liquid crystal approach is much better than previous soft-matter techniques, which had typically involved using intense light fields to change material properties of the soft matter, such as the index of refraction. The new method reduces the energy needed by more than a factor of a hundred. Because the STED laser pulse circulates repeatedly in the crystal, a single photon can deplete many dye molecules of the energy from the first laser pulse.</span> </p><p><a href="https://ravnik.fmf.uni-lj.si/" target="_blank">Miha Ravnik</a>, a theoretical physicist also at the University of Ljubljana who worked on the project, explains that control of light by light is essential in soft-matter photonic logic gates. “You can very much control when [light] is generated and in which direction,” Ravnik says of the light shined into the polymer waveguides. “And this gives you, then, this capability that you create logical operations with light.”</p><p>Aside from its potential in photonic logical circuits, the team’s approach presents several technical advantages over photonics made from silicon or other hard materials, Muševič says. For example, using soft matter greatly simplifies the manufacturing process. The liquid crystal in the team’s device can be inserted in less than a second, but manufacturing a similar structure with hard materials is difficult. Additionally, soft-matter devices can be manufactured at much lower temperatures than silicon and other hard materials. Muševič also points out that soft matter presents an opportunity to experiment with the geometry of the device. With liquid crystals “you can make many different kinds of cavities,” says Muševič. “You have, I would say, a lot of engineering space.”</p><p>Ravnik is excited for the potential of the team’s breakthrough, particularly as a step toward <a href="https://spectrum.ieee.org/generative-optical-ai-nature-ucla" target="_self">photonic computing</a> and even photonic neural networks. But, he recognizes that these developments are far down the line. “There’s no way this technology can compete with current neural network implementation at all,” he admits. Still, the possibilities are tantalizing. “The energy losses are predicted to be extremely low, the speeds for calculation extremely high.”</p>]]></description><pubDate>Mon, 13 Apr 2026 12:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/soft-photonics</guid><category>Flexible-circuits</category><category>Photonics</category><category>Optical-switch</category><dc:creator>Velvet Wu</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/illustration-of-a-micropipette-piercing-through-a-hemisphere-shaped-membrane-to-inject-a-droplet-at-its-core.jpg?id=65506297&amp;width=980"></media:content></item><item><title>HIPPO Turns One Master Password Into Many Without Storing Any</title><link>https://spectrum.ieee.org/storeless-password-manager</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/personified-dots-sneaking-out-of-a-hidden-password-field-on-a-login-page.jpg?id=65493656&width=1245&height=700&coordinates=0%2C62%2C0%2C63"/><br/><br/><p>
<em>This article is part of our exclusive <a href="https://spectrum.ieee.org/collections/journal-watch/" target="_blank">IEEE Journal Watch series</a> in partnership with IEEE Xplore.</em>
</p><p>Most people are all too familiar with attempting to type out a password multiple times—only to get locked out of their accounts, triggering a vicious cycle of new passwords that are quickly forgotten. <a data-linked-post="2650276006" href="https://spectrum.ieee.org/qa-paul-grassi-of-nist-on-what-makes-a-strong-password" target="_blank">Password managers</a> can be a helpful solution to sidestep this issue but also come with some risk if the saved passwords become compromised.</p><p>However, there is a different solution that does not involve saving passwords on a server. Instead, it requires a single master password that is easily remembered or written down on paper.</p><p>In a recent study, researchers found that people are willing to complete one extra step to access their accounts using the approach, which they report feeling is more secure and easier to use than traditional manual password entry. The <a href="https://ieeexplore.ieee.org/document/11415666" rel="noopener noreferrer" target="_blank">results</a> were published 27 February in <a href="https://ieeexplore.ieee.org/xpl/RecentIssue.jsp?punumber=4236" rel="noopener noreferrer" target="_blank"><em><em>IEEE Internet Computing</em></em></a>.</p><h2>HIPPO Password Manager Security</h2><p>Remembering multiple passwords across various accounts is a challenge for many people. Password managers avoid the need to memorize every single password by storing encrypted passwords in a secure online “vault.” Still, these vaults can be hacked. Malicious attackers can break into the vaults’ servers or steal the passwords by hacking the user’s own computer.</p><p>To overcome these challenges, a team of researchers created a password manager that doesn’t store the passwords, called HIPPO (Hidden-Password Online Password), which works like a browser extension.</p><p>The user needs to remember or write down only one master password. As they visit each site that requires them to log in, they enter their master password, and then HIPPO generates a site-specific password on the spot.</p><p>To do so, HIPPO’s browser extension first applies a <a data-linked-post="2650267393" href="https://spectrum.ieee.org/new-king-of-security-algorithms-crowned" target="_blank">cryptographic function</a> called an <a href="https://en.wikipedia.org/wiki/Oblivious_pseudorandom_function" rel="noopener noreferrer" target="_blank">oblivious pseudorandom function</a> to the master password. The result is a website-specific “masked” password that is sent to the HIPPO server. That server applies its own secret cryptographic key and sends the result back. The browser then removes its temporary mask and uses the result to generate the site-specific password on the spot. </p><p>“You can think of it as a calculator computing the exact same complex password on the spot every time you visit the site, eliminating the need to save it anywhere,” says <a href="https://sa.linkedin.com/in/mohammed-jubur-302495105" rel="noopener noreferrer" target="_blank">Mohammed Jubur</a>, an assistant professor of computer science at Jazan University, in Saudi Arabia, and co-creator of HIPPO. “Neither the master secret nor the derived site password is ever stored locally or remotely—the fresh-derived [password] is simply autofilled into the target site’s login field.”</p><h2>User Perception and Trust for Password Managers</h2><p>In their study, Jubur and colleagues had 25 volunteers provide feedback on HIPPO, comparing it to traditional manual password entry. Participants were tasked with setting up their accounts with a password given to them on a piece of paper. They were then asked to log in to a site 10 times using traditional manual password entry, and another 10 times using HIPPO. Participants rated their experience with each approach, evaluating different factors such as perceived security and ease of use.</p><p>The results show that users rated both approaches with a “good” usability score. However, Jubur notes, “participants perceived HIPPO to be significantly more secure and trustworthy compared to traditional password-only authentication.”</p><p>Whereas HIPPO received an average score of 4.04 out of 5 for perceived security, traditional manual password entry received a score of 3.09. Users also reported higher trust scores for HIPPO, at 4.00, compared to 3.30 for traditional password entry.</p><p>The researchers were surprised to learn that users also reported HIPPO as being easier to use—even though it requires an extra activation step, such as pressing F2 or entering a prefix like “@@” to activate the password-generation mode.</p><p>“We initially expected HIPPO’s usability to be merely comparable [to traditional password entry],” explains <a href="https://engineering.tamu.edu/cse/profiles/Saxena-Nitesh.html" rel="noopener noreferrer" target="_blank">Nitesh Saxena</a>, a professor in the Department of Computer Science and Engineering and associate director of the Global Cyber Research Institute (GCRI) at Texas A&M University, who co-created HIPPO with Jubur. “However, participants found the cognitive burden of repeatedly typing a complex random password to be so substantial, that even a tool with an extra step improved their experience.”</p><p>The researchers note that this was a small-scale, single-session study, so a follow-up study over a longer period is needed to explore HIPPO’s performance.</p><p>Jubur adds that, in future work, the team plans to evaluate longer-term life-cycle events, such as measuring the completion time, error rates, and lockout risk associated with HIPPO.</p><p>For example, he says, “we also plan to evaluate the user experience and the risk of account lockouts when a user needs to change their master password, which forces them to update their credentials across all their connected websites.”</p>]]></description><pubDate>Sat, 11 Apr 2026 13:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/storeless-password-manager</guid><category>Cryptography</category><category>Passwords</category><category>Journal-watch</category><dc:creator>Michelle Hampson</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/personified-dots-sneaking-out-of-a-hidden-password-field-on-a-login-page.jpg?id=65493656&amp;width=980"></media:content></item><item><title>Chip Can Project Video the Size of a Grain of Sand</title><link>https://spectrum.ieee.org/mems-photonics</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/an-array-of-tiny-metallic-cantilevers-curving-away-from-the-surface-of-a-photonic-chip.jpg?id=65493217&width=1245&height=700&coordinates=0%2C156%2C0%2C157"/><br/><br/><p><span>By many estimates, quantum computers will need <a href="https://spectrum.ieee.org/neutral-atom-quantum-computing" target="_blank">millions of qubits </a>to realize their potential applications in cybersecurity, drug development, and other industries. The problem is, anyone who has wanted to simultaneously control millions of a certain kind of qubit has run into the problem of trying to control millions of laser beams. </span> </p><p><span>That’s exactly the challenge that was faced by scientists working on the <a href="https://www.mitre.org/resources/quantum-moonshot" target="_blank">MITRE Quantum Moonshot project</a>, which brought together scientists from MITRE, MIT, the University of Colorado at Boulder, and Sandia National Laboratories. The solution they developed came in the form of an image projection technology that they realized could also be the fix for a host of other challenges in augmented reality, biomedical imaging, and elsewhere. The device is a 1-square-millimeter photonic chip capable of projecting the Mona Lisa onto an area smaller than the size of two human egg <a href="https://spectrum.ieee.org/embryo-electrode-array" target="_blank">cells</a>. </span> </p><p><span>“When we started, we certainly never would have anticipated that we would be making a technology that might revolutionize imaging,” says Matt Eichenfield, one of the leaders of the Quantum Moonshot project, a collaborative research effort focused on developing a scalable, diamond-based quantum computer, and a professor of quantum engineering at the University of Colorado at Boulder. Each second, their chip is capable of projecting 68.6 million individual spots of light—called scannable pixels—to differentiate them from physical pixels. That’s more than 50 times the capability of previous technology, such as <a href="https://spectrum.ieee.org/mems-lidar" target="_blank">micro-electromechanical systems (MEMS) micromirror arrays</a>.</span></p><p> <span>“We have now made a scannable pixel that is at the absolute limit of what diffraction allows,” says <a href="https://www.linkedin.com/in/y-henry-wen-2b41979/" target="_blank">Henry Wen</a>, a visiting researcher at MIT and a photonics engineer at <a href="https://www.quera.com/" target="_blank">QuEra Computing</a>.</span></p><p>The chip’s distinguishing feature is an array of tiny microscale cantilevers, which curve away from the plane of the chip in response to voltage and act as miniature “ski jumps” for light. Light is channeled along the length of each cantilever via a waveguide and exits at its tip. The cantilevers contain a thin layer of aluminum nitride, a piezoelectric that expands or contracts under voltage, thus moving the micromachine up and down and enabling the array to scan beams of light over a two-dimensional area.</p><p>Despite the magnitude of the team’s achievement, Eichenfield says that the process of engineering the cantilevers was “pretty smooth.” Each cantilever is composed of a stack of several submicrometer layers of material and curls approximately 90 degrees out of the plane at rest. To achieve such a high curvature, the team took advantage of differences in the contraction and expansion of individual layers caused by physical stresses in the material resulting from the fabrication process. The materials are first deposited flat onto the chip. Then, a layer in the chip below the cantilever is removed, allowing the material stresses to take effect, releasing the cantilever from the chip and allowing it to curl out. The top layer of each cantilever also features a series of silicon dioxide bars running perpendicular to the waveguide, which keep the cantilever from curling along its width while also improving its lengthwise curvature.</p><p class="shortcode-media shortcode-media-youtube"> <span class="rm-shortcode" data-rm-shortcode-id="5525c992b93704c6dfdada2cd2c1d9c2" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/A4-ZqQTZauw?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span> <small class="image-media media-caption" placeholder="Add Photo Caption...">A micro-cantilever wiggles and waggles to project light in the right place.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Matt Saha, Y. Henry Wen, et al.</small></p><p>What was more of a challenge than engineering the chip itself was figuring out the details of actually making the chip project images and videos. Working out the process of synchronizing and timing the cantilevers’ motion and light beams to generate the right colors at the right time was a substantial effort, according to <a href="https://www.linkedin.com/in/agreenspon/" target="_blank">Andy Greenspon</a>, a researcher at MITRE who also worked on the project. Now, the team has successfully projected a variety of videos from a single cantilever, including clips from the movie <em><em><a href="https://www.youtube.com/watch?v=GPG3zSgm_Qo&list=PLnvfBuirq7alZgA0yGBnNObE5CeJTpUW4" target="_blank">A Charlie Brown Christmas</a></em></em>. </p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="A warped projection of the Mona Lisa." class="rm-shortcode" data-rm-shortcode-id="a4e5294e1a010872e545dbc18fb0e208" data-rm-shortcode-name="rebelmouse-image" id="a1039" loading="lazy" src="https://spectrum.ieee.org/media-library/a-warped-projection-of-the-mona-lisa.jpg?id=65493253&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">The chip projected a roughly 125-micrometer image of the Mona Lisa.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit..."><a href="https://www.nature.com/articles/s41586-025-10038-6" target="_blank">Matt Saha, Y. Henry Wen, et al.</a></small></p><p>Because the chip can project so many more spots in any given time interval than any previous beam scanners, it could also be used to control many more qubits in quantum computers. The Quantum Moonshot program’s mission is to build a quantum computer that can be scaled to millions of qubits. So clearly, it needs a scalable way of controlling each one, explains Wen. Instead of using one laser per qubit, the team realized that not every qubit needed to be controlled at every given moment. The chip’s ability to move light beams over a two-dimensional area would allow them to control all of the qubits with many fewer lasers. </p><p>Another process that Wen thinks the chip could improve is scanning objects for <a href="https://spectrum.ieee.org/3d-printed-linear-motor" target="_blank">3D printing</a>. Today, that typically involves using a single laser to scan over the entire surface of an object. The new chip, however, could potentially employ thousands of laser beams. “I think now you can take a process that would have taken hours and maybe bring it down to minutes,” says Wen. </p><p>Wen is also excited to explore the potential of different cantilever shapes. By changing the orientations of the bars perpendicular to the waveguide, the team has been able to make the cantilevers curl into helixes. Wen says that such unusual shapes could be useful in making a <a href="https://spectrum.ieee.org/neurobot-living-robot-nervous-system" target="_blank">lab-on-a-chip for cell biology</a> or <a href="https://spectrum.ieee.org/lab-on-a-chip-grippers" target="_blank">drug development</a>. “A lot of this stuff is imaging, scanning a laser across something, either to image it or to stimulate some response. And so we could have one of these ski jumps curl not just up, but actually curl back around, and then move around and scan over a sample,” Wen explains. “If you can imagine a structure that will be useful for you, we should try it.”</p>]]></description><pubDate>Thu, 09 Apr 2026 13:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/mems-photonics</guid><category>Microarray</category><category>Digital-micromirror-device</category><category>Mems</category><category>Quantum-computers</category><category>Nitrogen-vacancy-defects-diamond</category><dc:creator>Velvet Wu</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/an-array-of-tiny-metallic-cantilevers-curving-away-from-the-surface-of-a-photonic-chip.jpg?id=65493217&amp;width=980"></media:content></item><item><title>AI Models Trained on Physics Are Changing Engineering</title><link>https://spectrum.ieee.org/large-physics-models-design-engineering</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/diagram-of-airflow-over-a-moving-sedan.jpg?id=65494121&width=1245&height=700&coordinates=0%2C62%2C0%2C63"/><br/><br/><p>Large language models have already <a href="https://spectrum.ieee.org/best-ai-coding-tools" target="_self">transformed</a> software engineering, for better or worse. Now, so-called large physics models are also starting to transform design engineering. These tools are beginning to replace—or at least amend—the role of full-fledged physics simulation in the automotive and aerospace industries, semiconductor engineering, and more.</p><p>Before the advent of computer simulation, a car manufacturer, for example, would create prototypes to test their designs, says <a href="https://www.linkedin.com/in/thomas-von-tschammer/" rel="noopener noreferrer" target="_blank">Thomas von Tschammer</a>, managing director at physics-based AI company <a href="https://www.neuralconcept.com/" rel="noopener noreferrer" target="_blank">Neural Concept</a>. “For the past 40 years, we reduced a lot of the need for prototypes by using numerical simulations for aerodynamics, for crash testing, and so on.” Now, von Tschammer explains, AI is drastically reducing the need for simulation, the same way simulation reduced the need for physical prototypes.</p><p>Growing adoptions of this type of AI was a topic of interest at <a href="https://www.nvidia.com/gtc/" rel="noopener noreferrer" target="_blank">Nvidia GTC</a> in March. <a href="https://www.linkedin.com/in/chris-johnston-/" rel="noopener noreferrer" target="_blank">Chris Johnston</a>, senior technical specialist at Jaguar Land Rover, <a href="https://www.nvidia.com/en-us/on-demand/session/gtc26-s81736/?playlistId=gtc26-industrial-engineering" rel="noopener noreferrer" target="_blank">presented</a> how his company is using Neural Concept’s technology. <a href="https://www.physicsx.ai/" rel="noopener noreferrer" target="_blank">PhysicsX</a>, another physics-based AI company, <a href="https://www.physicsx.ai/newsroom/physicsx-announces-advancement-to-open-standards-for-physics-ai-powered-by-nvidia" rel="noopener noreferrer" target="_blank">announced</a> a collaboration with Nvidia to advance open standards for such models, also at GTC.</p><h2>The AI design engineering workflow</h2><p>Over the past six months, <a href="https://www.gm.com/" rel="noopener noreferrer" target="_blank">General Motors</a> (GM) has introduced large physics models into their car design process to speed up the workflow. </p><p>Previously, a creative design engineer would develop a 3D model of a new car concept. This model would be sent to aerodynamics specialists, who would run physics simulations to determine the coefficient of drag of the proposed car—an important metric for energy efficiency of the vehicle. This simulation phase would take about two weeks, and the aerodynamics engineer would then report the drag coefficient back to the creative designer, possibly with suggested modifications.</p><p>Now, GM has trained an in-house large physics model on those simulation results. The AI takes in a 3D car model and outputs a coefficient of drag in a matter of minutes. “We have experts in the aerodynamics and the creative studio now who can sit together and iterate instantly to make decisions [about] our future products,” says <a href="https://www.linkedin.com/in/rdstrauss/" rel="noopener noreferrer" target="_blank">Rene Strauss</a>, director of virtual integration engineering at GM. </p><p>For GM and other companies, running inference on an AI model trained on physics simulations, instead of running the simulation itself, can bring immense time savings. “Depending on the kinds of physics [being simulated], or the resolution, it can be anywhere between 10,000 to close to a million times faster,” says <a href="https://www.linkedin.com/in/jacomo-corbo/" rel="noopener noreferrer" target="_blank">Jacomo Corbo</a>, CEO and co-founder of PhysicsX.</p><h2>How accurate are large physics models?</h2><p>But what about accuracy? For GM’s purposes, Strauss says accuracy is not a huge concern at the design stage because finer details are ironed out later in the process. “When it really starts to matter is when we’re getting close to launching a vehicle, and the coefficient of drag is going to be used for our energy calculation, which eventually goes to the certification of our miles per gallon on the sticker.” At that stage, Strauss says, a physical model of the car will be put into a wind tunnel for an exact number.</p><p>PhysicsX’s Corbo argues that, with the right data, the AI model accuracy can supersede the accuracy of the simulation it’s trained on. The trick is to incorporate experimental measurements to fine-tune the model. If a physics simulation doesn’t agree exactly with experimental data, it is often difficult to figure out why and tweak the model until they agree. With AI, incorporating a few experimental examples into the training process is a lot more straightforward, and it’s not necessary to understand where exactly the model went wrong.</p><p>All in all, by drastically bringing down the time it takes to model the physics, large physics models enable engineers to explore a much greater range of possibilities before a final design is reached. </p><h2>Training large physics models</h2><p>There is no one-size-fits-all approach to training large physics models. Depending on the types of data available, and the physics in question, the models may use the <a href="https://spectrum.ieee.org/what-is-generative-ai" target="_self">transformer</a> architecture that underlies LLMs, a generalized version of convolutional neural networks known as <a href="https://dataroots.io/blog/a-gentle-introduction-to-geometric" rel="noopener noreferrer" target="_blank">geometric deep learning</a>, or an architecture that can solve partial differential equations called <a href="https://zongyi-li.github.io/neural-operator/" rel="noopener noreferrer" target="_blank">neural operators</a>.</p><p>Currently, most companies are training their own models on their simulation data, catering to specific use cases. In GM’s aerodynamics implementation, there are different AI models for different types of cars: think SUVs versus sedans. But PhysicsX’s Corbo says his team is working on building more “foundational” physics models that can be applied across different scenarios.</p><p>Both <a href="https://arxiv.org/pdf/2001.08361" rel="noopener noreferrer" target="_blank">LLMs</a> and <a href="https://spectrum.ieee.org/solve-robotics" target="_self">robotics</a> have benefited from scaling laws, which describe how a system improves as the models increase in size or get trained on more data. In AI, models tend to improve quickly, in a nonlinear way. Along the way, the models also become more generalizable—extending them to new settings takes less and less fine-tuning to reach the same accuracy. Corbo says his team is now starting to see the same types of scaling laws for large physics models.</p><p>“What we’re seeing here is maybe a little bit unsurprising,” Corbo says, “but it’s also pretty incredible. And it’s given us the confidence to make these models bigger, because they perform a whole lot better, and they cover broader domains, and they have these really amazing emergent properties.”</p><p>Developing open standards for the data formats used in training, as well as the model architectures, should help develop these more powerful foundational models. That’s the goal of PhysicsX’s collaboration with Nvidia, and of Nvidia’s <a href="https://developer.nvidia.com/blog/physics-ml-platform-physicsnemo-is-now-open-source/" rel="noopener noreferrer" target="_blank">physicsNeMo</a> open source platform.</p><p>“The thing that we’re collaborating on is being able to compose architectures from building blocks,” Corbo says, making it easy for those in both academia and industry to reuse and build upon existing models.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="A physics-based AI software being used to generate a 3D geometry of a data center server PCB ready to be run in computational fluid dynamics." class="rm-shortcode" data-rm-shortcode-id="617060afa65951af85acad5c9b9c8708" data-rm-shortcode-name="rebelmouse-image" id="70269" loading="lazy" src="https://spectrum.ieee.org/media-library/a-physics-based-ai-software-being-used-to-generate-a-3d-geometry-of-a-data-center-server-pcb-ready-to-be-run-in-computational-fl.jpg?id=65494125&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">A type of AI called a large physics model is used by an engineer to quickly generate heat flow in a 3D data center server design. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Neural Concept</small></p><h2>The long-term role of simulations and engineers</h2><p>While some are working on developing more powerful models, others are pushing to implement what’s already available into existing workflows, which is no easy task. “With any innovation, it’s not a straight line. There’s some steps forward and then some steps back and improvements that we find along the way. But that’s part of the joy of the innovation process and using new tools like this,” GM’s Strauss says.</p><p>This technology is still in the early stages, and it’s unclear what the final role of AI tools will be in the engineering workflow. For one, opinions vary on whether AI will replace simulations completely, or just reduce their use.</p><p>“We will never fully replace simulations,” Neural Concept’s von Tschammer says. “But the idea is to make a much smarter usage of simulation at the most major phase of developments, and you use AI to speed up the early design stages, where you need to explore a very wide set of options.”</p><p>PhysicsX’s Corbo begs to differ. “The whole idea is to take numerical simulation … out of the workflow,” he says, “and to move that to inference.”</p><p>Whatever the role of simulation will be, everyone in the field is adamant that human design engineers will continue to be in the driver’s seat, enabled by these newfangled tools  to do their best work. (After all, when has AI ever threatened to replace human labor?)</p><p>“What we’re seeing is that actually, these tools are empowering the engineers to be much more efficient,” von Tschammer says. “Before, these engineers would spend a lot of time on low-added-value tasks, whereas now these manual tasks from the past can be automated using these AI models, and the engineers can focus on taking the design decisions at the end of the day. We still need engineers more than ever.”</p>]]></description><pubDate>Thu, 09 Apr 2026 11:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/large-physics-models-design-engineering</guid><category>Physics-simulations</category><category>General-motors</category><category>Nvidia-gtc</category><category>Engineering-design</category><dc:creator>Dina Genkina</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/diagram-of-airflow-over-a-moving-sedan.jpg?id=65494121&amp;width=980"></media:content></item><item><title>Decentralized Training Can Help Solve AI’s Energy Woes</title><link>https://spectrum.ieee.org/decentralized-ai-training-2676670858</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/illustration-of-several-data-servers-interconnected-across-long-distances.jpg?id=65477795&width=1245&height=700&coordinates=0%2C156%2C0%2C157"/><br/><br/><p> <a href="https://spectrum.ieee.org/topic/artificial-intelligence/" target="_self">Artificial intelligence</a> harbors an enormous <a href="https://spectrum.ieee.org/topic/energy/" target="_self">energy</a> appetite. Such constant cravings are evident in the <a href="https://spectrum.ieee.org/ai-index-2025" target="_self">hefty carbon footprint</a> of the <a href="https://spectrum.ieee.org/tag/data-centers" target="_self">data centers</a> behind the AI boom and the steady increase over time of <a href="https://spectrum.ieee.org/tag/carbon-emissions" target="_self">carbon emissions</a> from training frontier <a href="https://spectrum.ieee.org/tag/ai-models" target="_self">AI models</a>.</p><p>No wonder big tech companies are warming up to <a href="https://spectrum.ieee.org/tag/nuclear-energy" target="_self">nuclear energy</a>, envisioning a future fueled by reliable, carbon-free sources. But while <a href="https://spectrum.ieee.org/nuclear-powered-data-center" target="_self">nuclear-powered data centers</a> might still be years away, some in the research and industry spheres are taking action right now to curb AI’s growing energy demands. They’re tackling training as one of the most energy-intensive phases in a model’s life cycle, focusing their efforts on decentralization.</p><p>Decentralization allocates model training across a network of independent nodes rather than relying on one platform or provider. It allows compute to go where the energy is—be it a dormant server sitting in a research lab or a computer in a <a href="https://spectrum.ieee.org/tag/solar-power" target="_self">solar-powered</a> home. Instead of constructing more data centers that require <a href="https://spectrum.ieee.org/tag/power-grid" target="_self">electric grids</a> to scale up their infrastructure and capacity, decentralization harnesses energy from existing sources, avoiding adding more power into the mix.</p><h2>Hardware in harmony</h2><p>Training AI models is a huge data center sport, synchronized across clusters of closely connected <a href="https://spectrum.ieee.org/tag/gpus" target="_self">GPUs</a>. But as <a href="https://spectrum.ieee.org/mlperf-trends" target="_self">hardware improvements struggle to keep up</a> with the swift rise in the size of <a href="https://spectrum.ieee.org/tag/large-language-models" target="_self">large language models</a>, even massive single data centers are no longer cutting it.</p><p>Tech firms are turning to the pooled power of multiple data centers—no matter their location. <a href="https://spectrum.ieee.org/tag/nvidia" target="_self">Nvidia</a>, for instance, launched the <a href="https://developer.nvidia.com/blog/how-to-connect-distributed-data-centers-into-large-ai-factories-with-scale-across-networking/" target="_blank">Spectrum-XGS Ethernet for scale-across networking</a>, which “can deliver the performance needed for large-scale single job AI training and inference across geographically separated data centers.” Similarly, <a href="https://spectrum.ieee.org/tag/cisco" target="_self">Cisco</a> introduced its <a href="https://blogs.cisco.com/sp/the-new-benchmark-for-distributed-ai-networking" target="_blank">8223 router</a> designed to “connect geographically dispersed AI clusters.”</p><p>Other companies are harvesting idle compute in <a href="https://spectrum.ieee.org/tag/servers" target="_self">servers</a>, sparking the emergence of a <a href="https://spectrum.ieee.org/gpu-as-a-service" target="_self">GPU-as-a-Service</a> business model. Take <a href="https://akash.network/" rel="noopener noreferrer" target="_blank">Akash Network</a>, a peer-to-peer <a href="https://spectrum.ieee.org/tag/cloud-computing" target="_self">cloud computing</a> marketplace that bills itself as the “Airbnb for data centers.” Those with unused or underused GPUs in offices and smaller data centers register as providers, while those in need of computing power are considered as tenants who can choose among providers and rent their GPUs.</p><p>“If you look at [AI] training today, it’s very dependent on the latest and greatest GPUs,” says Akash cofounder and CEO <a href="https://www.linkedin.com/in/gosuri" rel="noopener noreferrer" target="_blank">Greg Osuri</a>. “The world is transitioning, fortunately, from only relying on large, high-density GPUs to now considering smaller GPUs.”</p><h2>Software in sync</h2><p>In addition to orchestrating the <a href="https://spectrum.ieee.org/tag/hardware" target="_self">hardware</a>, decentralized AI training also requires algorithmic changes on the <a href="https://spectrum.ieee.org/tag/software" target="_self">software</a> side. This is where <a href="https://cloud.google.com/discover/what-is-federated-learning" rel="noopener noreferrer" target="_blank">federated learning</a>, a form of distributed <a href="https://spectrum.ieee.org/tag/machine-learning" target="_self">machine learning</a>, comes in.</p><p>It starts with an initial version of a global AI model housed in a trusted entity such as a central server. The server distributes the model to participating organizations, which train it locally on their data and share only the model weights with the trusted entity, explains <a href="https://www.csail.mit.edu/person/lalana-kagal" rel="noopener noreferrer" target="_blank">Lalana Kagal</a>, a principal research scientist at <a href="https://www.csail.mit.edu/" rel="noopener noreferrer" target="_blank">MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL)</a> who leads the <a href="https://www.csail.mit.edu/research/decentralized-information-group-dig" rel="noopener noreferrer" target="_blank">Decentralized Information Group</a>. The trusted entity then aggregates the weights, often by averaging them, integrates them into the global model, and sends the updated model back to the participants. This collaborative training cycle repeats until the model is considered fully trained.</p><p>But there are drawbacks to distributing both data and computation. The constant back-and-forth exchanges of model weights, for instance, result in high communication costs. Fault tolerance is another issue.</p><p>“A big thing about AI is that every training step is not fault-tolerant,” Osuri says. “That means if one node goes down, you have to restore the whole batch again.”</p><p>To overcome these hurdles, researchers at <a href="https://deepmind.google/" rel="noopener noreferrer" target="_blank">Google DeepMind</a> developed <a href="https://arxiv.org/abs/2311.08105" rel="noopener noreferrer" target="_blank">DiLoCo</a>, a distributed low-communication optimization <a href="https://spectrum.ieee.org/tag/algorithms" target="_self">algorithm</a>. DiLoCo forms what <a href="https://spectrum.ieee.org/tag/google-deepmind" target="_self">Google DeepMind</a> research scientist <a href="https://arthurdouillard.com/" rel="noopener noreferrer" target="_blank">Arthur Douillard</a> calls “islands of compute,” where each island consists of a group of <a href="https://spectrum.ieee.org/tag/chips" target="_self">chips</a>. Every island holds a different chip type, but chips within an island must be of the same type. Islands are decoupled from each other, and synchronizing knowledge between them happens once in a while. This decoupling means islands can perform training steps independently without communicating as often, and chips can fail without having to interrupt the remaining healthy chips. However, the team’s experiments found diminishing performance after eight islands.</p><p>An improved version, dubbed <a href="https://arxiv.org/abs/2501.18512" rel="noopener noreferrer" target="_blank">Streaming DiLoCo</a>, further reduces the bandwidth requirement by synchronizing knowledge “in a streaming fashion across several steps and without stopping for communicating,” says Douillard. The mechanism is akin to watching a video even if it hasn’t been fully downloaded yet. “In Streaming DiLoCo, as you do computational work, the knowledge is being synchronized gradually in the background,” he adds.</p><p>AI development platform <a href="https://www.primeintellect.ai/" rel="noopener noreferrer" target="_blank">Prime Intellect</a> implemented a variant of the DiLoCo algorithm as a vital component of its 10-billion-parameter <a href="https://www.primeintellect.ai/blog/intellect-1-release" rel="noopener noreferrer" target="_blank">INTELLECT-1</a> model trained across five countries spanning three continents. Upping the ante, <a href="https://0g.ai/" rel="noopener noreferrer" target="_blank">0G Labs</a>, makers of a decentralized AI <a href="https://spectrum.ieee.org/tag/operating-system" target="_self">operating system</a>, <a href="https://0g.ai/blog/worlds-first-distributed-100b-parameter-ai" rel="noopener noreferrer" target="_blank">adapted DiLoCo to train a 107-billion-parameter foundation model</a> under a network of segregated clusters with limited bandwidth. Meanwhile, popular <a href="https://spectrum.ieee.org/tag/open-source" target="_self">open-source</a> <a href="https://spectrum.ieee.org/tag/deep-learning" target="_self">deep learning</a> framework <a href="https://pytorch.org/projects/pytorch/" rel="noopener noreferrer" target="_blank">PyTorch</a> included DiLoCo in its <a href="https://meta-pytorch.org/torchft/" rel="noopener noreferrer" target="_blank">repository of fault-tolerance techniques</a>.</p><p>“A lot of engineering has been done by the community to take our DiLoCo paper and integrate it in a system learning over consumer-grade internet,” Douillard says. “I’m very excited to see my research being useful.”</p><h2>A more energy-efficient way to train AI</h2><p>With hardware and software enhancements in place, decentralized AI training is primed to help solve AI’s energy problem. This approach offers the option of training models “in a cheaper, more resource-efficient, more energy-efficient way,” says MIT CSAIL’s Kagal.</p><p>And while Douillard admits that “training methods like DiLoCo are arguably more complex, they provide an interesting trade-off of system efficiency.” For instance, you can now use data centers across far apart locations without needing to build ultrafast bandwidth in between. Douillard adds that fault tolerance is baked in because “the blast radius of a chip failing is limited to its island of compute.”</p><p>Even better, companies can take advantage of existing underutilized processing capacity rather than continuously building new energy-hungry data centers. Betting big on such an opportunity, Akash created its <a href="https://www.youtube.com/watch?v=zAj41xSNPeI" rel="noopener noreferrer" target="_blank">Starcluster program</a>. One of the program’s aims involves tapping into solar-powered homes and employing the desktops and laptops within them to train AI models. “We want to convert your home into a fully functional data center,” Osuri says.</p><p>Osuri acknowledges that participating in Starcluster will not be trivial. Beyond solar panels and devices equipped with consumer-grade GPUs, participants would also need to invest in <a href="https://spectrum.ieee.org/tag/batteries" target="_self">batteries</a> for backup power and redundant internet to prevent downtime. The Starcluster program is figuring out ways to package all these aspects together and make it easier for homeowners, including collaborating with industry partners to subsidize battery costs.</p><p>Back-end work is already underway to enable <a href="https://akash.network/roadmap/aep-60/" rel="noopener noreferrer" target="_blank">homes to participate as providers in the Akash Network</a>, and the team hopes to reach its target by 2027. The Starcluster program also envisions expanding into other solar-powered locations, such as schools and local community sites.</p><p>Decentralized AI training holds much promise to steer AI toward a more environmentally sustainable future. For Osuri, such potential lies in moving AI “to where the energy is instead of moving the energy to where AI is.”</p>]]></description><pubDate>Tue, 07 Apr 2026 14:00:01 +0000</pubDate><guid>https://spectrum.ieee.org/decentralized-ai-training-2676670858</guid><category>Training</category><category>Ai-energy</category><category>Data-center</category><category>Large-language-models</category><dc:creator>Rina Diane Caballar</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/illustration-of-several-data-servers-interconnected-across-long-distances.jpg?id=65477795&amp;width=980"></media:content></item><item><title>ENIAC’s Architects Wove Stories Through Computing</title><link>https://spectrum.ieee.org/eniac-80th-anniversary-weaving</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/close-up-black-and-white-1940-s-image-of-a-woman-holding-a-metallic-brick-like-controller-with-large-knobs.jpg?id=65453792&width=1245&height=700&coordinates=0%2C187%2C0%2C188"/><br/><br/><p><em><em>This year marks the </em></em><a href="https://spectrum.ieee.org/eniac-80-ieee-milestone" target="_self"><em><em>80th anniversary of ENIAC</em></em></a><em><em>, the first general-purpose digital computer. The computer was built during World War II to speed up ballistics calculations, but its contributions to computing extend well beyond military applications.</em></em></p><div class="rm-embed embed-media"><iframe height="110px" id="noa-web-audio-player" src="https://embed-player.newsoveraudio.com/v4?key=q5m19e&id=https://spectrum.ieee.org/eniac-80th-anniversary-weaving&bgColor=F5F5F5&color=1b1b1c&playColor=1b1b1c&progressBgColor=F5F5F5&progressBorderColor=bdbbbb&titleColor=1b1b1c&timeColor=1b1b1c&speedColor=1b1b1c&noaLinkColor=556B7D&noaLinkHighlightColor=FF4B00&feedbackButton=true" style="border: none" width="100%"></iframe></div><p><em><em>Two of ENIAC’s key architects—John W. Mauchly, its co-inventor, and Kathleen “Kay” McNulty, one of the <a href="https://spectrum.ieee.org/eniac-woman-programmers" target="_blank">six original programmers</a>—married a few years after its completion and raised seven children together. Mauchly and McNulty’s grandchild Naomi Most </em></em><a href="https://youtu.be/XYEVmqGhVxo?si=fseDLKFz1W8meWR6&t=4515" rel="noopener noreferrer" target="_blank"><em><em>delivered a talk</em></em></a><em><em> as part of a celebration in honor of ENIAC’s anniversary on 15 February, which was held online and in-person at the American Helicopter Museum in West Chester, Pa. The following is adapted from that presentation.</em></em></p><p class="ieee-inbody-related">RELATED: <a href="https://spectrum.ieee.org/eniac-80-ieee-milestone" target="_blank">ENIAC, the First General-Purpose Digital Computer, Turns 80</a></p><p>There was a library at my grandparents’ farmhouse that felt like it went on forever. September light through the windows, beech leaves rustling outside on the stone porch, the sounds of cousins and aunts and uncles somewhere in the house. And in the corner of that library, an IBM personal computer.</p><p>When I spent summers there as a child, I didn’t yet know that the computer was closely tied to my family’s story.</p><p>My grandparents are known for their contributions to creating the Electronic Numerical Integrator and Computer, or ENIAC. But both were interested in more than just crunching numbers: My grandfather wanted to predict the weather. My grandmother wanted to be a good storyteller. </p><p>In Irish, the first language my grandmother Kathleen “Kay” McNulty ever spoke, a word existed to describe both of these impulses: <em><em>ríomh</em></em>.</p><p>I began to learn the Irish language myself five years ago, and I was struck by how certain words and phrases had multiple meanings. According to renowned Irish cultural historian Manchán Magan—from whom I took lessons—the word <em><em>ríomh</em></em> has at different times been used to mean to compute, but also <a href="https://www.making.ie/stories/irish-words-weaving" rel="noopener noreferrer" target="_blank">to weave, to narrate, or to compose a poem</a>. That one word that can tell the story of ENIAC, a machine with wires woven like thread that was built to compute, make predictions, and search for a signal in the noise. </p><h2>John Mauchly’s Weather-Prediction Ambitions</h2><p>Before working on ENIAC, John Mauchly <a href="https://fi.edu/en/news/case-files-john-w-mauchly-and-j-presper-eckert" rel="noopener noreferrer" target="_blank">spent years collecting rainfall data</a> across the United States. His favorite pastime was meteorology, and he wanted to find patterns in storm systems to predict the weather.</p><p>The Army, however, funded ENIAC to make simpler predictions: calculating ballistic trajectory tables. Start there, co-inventors J. Presper Eckert and Mauchly realized, and perhaps the weather would soon be computable.</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="Black and white 1960s image of two white men in suits looking at a wall of computer controls." class="rm-shortcode" data-rm-shortcode-id="7872d50df109149c936e400909defc38" data-rm-shortcode-name="rebelmouse-image" id="75108" loading="lazy" src="https://spectrum.ieee.org/media-library/black-and-white-1960s-image-of-two-white-men-in-suits-looking-at-a-wall-of-computer-controls.jpg?id=65428294&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Co-inventors John Mauchly [left] and J. Presper Eckert look at a portion of ENIAC on 25 November 1966. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Hulton Archive/Getty Images</small></p><p>Weather is a system unfolding through time, and a model of a storm is a story about how that system might unfold. There’s an old Irish saying related to this idea: <a href="https://daltai.com/is-maith-an-scealai-an-aimsir/" target="_blank"><em><em>Is maith an scéalaí an aimsir</em></em></a><em><em>.</em></em> Literally, “weather is a good storyteller.” But <em><em>aimsir</em></em> also means time. So the usual translation of this phrase into English becomes “time will tell.”</p><p>Mauchly wanted to <em><em>ríomh an aimsire</em></em>—to weave the weather into pattern, to compute the storm, to narrate the chaos. He realized that complex systems don’t reveal their full purpose at conception. They reveal it through <em><em>aimsir</em></em>—through weather, through time, through use.</p><h2>ENIAC’s First Programmers Were Weavers</h2><p>Kathleen “Kay” McNulty was born on 12 February 1921, in Creeslough, Ireland, on the night <a href="https://en.wikipedia.org/wiki/James_McNulty_(Irish_activist)" target="_blank">her father</a>—an IRA training officer—was arrested and imprisoned in Derry Gaol.</p><p>Family oral history holds that her people were weavers. She spoke only Irish until her family reached Philadelphia when she was 4 years old, entering American school the following year knowing virtually no English. She graduated in 1942 from Chestnut Hill College with a mathematics degree, was recruited to compute artillery firing tables by hand for the U.S. Army, and was then selected—along with <a href="https://spectrum.ieee.org/the-women-behind-eniac" target="_blank">five other women</a>—to program ENIAC.</p><p>They had no manual. They had only blueprints.</p><p>McNulty and her colleagues learned ENIAC and its quirks the way you learn a loom: by touch, by memory, by routing threads of electricity into patterns. They developed embodied knowledge the designers could only approximate. They could narrow a malfunction to a specific failed vacuum tube before any technician could locate it.</p><p>McNulty and Mauchly are also credited with conceiving the subroutine, the sequence of instructions that can be repeatedly recalled to perform a task, now essential in any programming. The subroutine was not in ENIAC’s blueprints, nor in the funding proposal. The concept emerged as highly determined people extended their imagination into the machine’s affordances.</p><p>The engineers designed the loom. Weavers discovered its true capabilities.</p><p>In 1950, four years after ENIAC was switched on, Mauchly’s dream was realized as it was used in the <a href="https://www.guinnessworldrecords.com/world-records/775520-first-computer-assisted-weather-forecast" target="_blank">world’s first computer-assisted weather forecast</a>. That was made possible after Klara von Neumann and Nick Metropolis reassembled and upgraded the ENIAC with a small amount of digital program memory. The programmers who transformed the math into operational code for the ENIAC were Norma Gilbarg, Ellen-Kristine Eliassen, and Margaret Smagorinsky. Their names are not as well-known as they should be.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Black and white 1940s image of three women operating a differential analyser in a basement." class="rm-shortcode" data-rm-shortcode-id="298168a77d38fd343eeb7d4bbfc219a7" data-rm-shortcode-name="rebelmouse-image" id="aacec" loading="lazy" src="https://spectrum.ieee.org/media-library/black-and-white-1940s-image-of-three-women-operating-a-differential-analyser-in-a-basement.jpg?id=65453828&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Before programming ENIAC, Kay McNulty [left] was recruited by the U.S. Army to compute artillery firing tables. Here, she and two other women, Alyse Snyder [center] and Sis Stump, operate a mechanical analog computer designed to solve differential equations in the basement of the University of Pennsylvania’s Moore School of Electrical Engineering.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">University of Pennsylvania</small></p><h2>Kay McNulty, Family Storyteller</h2><p>Kay married John Mauchly in 1948, describing him as “the greatest delight of my life. He was so intelligent and had so many ideas.... He was not only lovable, he was loving.” She spent the rest of her life ensuring he, Eckert, and the ENIAC programmers would be recognized.</p><p>When she died in 2006, I came to her funeral in shock, not fully knowing what I’d lost. As she drifted away, it was said, she had been reciting her prayers in Irish. This understanding made it quickly over to Creeslough, in County Donegal, and awaited me when I visited to honor her memory with the <a href="https://www.youtube.com/watch?v=zbkk2RJMW9g" target="_blank">dedication of a plaque</a> right there in the center of town.</p><p>In <a href="https://mathshistory.st-andrews.ac.uk/Extras/Mauchly_Antonelli_story" target="_blank">her own memoir</a>, she wrote: “If I am remembered at all, I would like to be remembered as my family storyteller.”</p><p>In Irish, the word for computer is <em><em>ríomhaire</em></em>. One who ríomhs. One who weaves, computes, and tells. My grandfather wanted to tell the story of the weather through computing. My grandmother wanted to be remembered as a storyteller. The language of her childhood already had a word that contained both of those ambitions.</p><h2>Computers as Narrative Engines</h2><p>When it was built, ENIAC looked like the back room of a textile production house. Panels. Switchboards. A room full of wires. Thread.</p><p>Thread does not tell you what it will become. We tend to think of computing as calculation—discrete and deterministic. But a model is a structured story about how something behaves.</p><p>Weather models, ballistic tables, economic forecasts, neural networks: These are all narrative engines, systems that take raw inputs and produce accounts of how the world might unfold. In complex systems, when parts are woven together through use, new structures arise that no one specified in advance.</p><p>Like ENIAC, the machines we are building now—the large models, the autonomous systems—are not merely calculators. They are looms.</p><p>Their most important properties will not be specified in advance. They will emerge through use, through the people who learn how to weave with them.</p><p>Through imagination.</p><p>Through <em><em>aimsir</em></em>.</p>]]></description><pubDate>Fri, 03 Apr 2026 13:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/eniac-80th-anniversary-weaving</guid><category>Eniac</category><category>Weather-prediction</category><category>Computer-history</category><category>Ireland</category><dc:creator>Naomi Most</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/close-up-black-and-white-1940-s-image-of-a-woman-holding-a-metallic-brick-like-controller-with-large-knobs.jpg?id=65453792&amp;width=980"></media:content></item><item><title>The AI Data Centers That Fit on a Truck</title><link>https://spectrum.ieee.org/modular-data-center</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/overhead-view-of-two-data-center-pods-each-measuring-55-feet-long-by-12-5-feet-wide.jpg?id=65417343&width=1245&height=700&coordinates=0%2C187%2C0%2C188"/><br/><br/><p>A <a data-linked-post="2676577917" href="https://spectrum.ieee.org/5gw-data-center" target="_blank">traditional</a> data center protects the expensive hardware inside it with a “shell” constructed from steel and concrete. Constructing a data center’s shell is inexpensive compared to the cost of the hardware and infrastructure inside it, but it’s not trivial. It takes time for engineers to consider potential sites, apply for permits, and coordinate with construction contractors.</p><p>That’s a problem for those looking to quickly deploy AI hardware, which has led companies like <a href="https://duosedge.ai/home" target="_blank">Duos Edge AI</a> and <a href="https://www.lgcns.com/en" target="_blank">LG CNS</a> to respond with a more modular approach. They use pre-fabricated, self-contained boxes that can be deployed in months instead of years. The boxes can operate alone or in tandem with others, providing the option to add more if required.</p><p>“I just came back from Nvidia’s GTC, and a lot of [companies] are sitting on their deployment because their data centers aren’t ready, or they can’t find the space,” said <a href="https://www.linkedin.com/in/doug-recker/" rel="noopener noreferrer" target="_blank">Doug Recker</a>, CEO of Duos Edge AI. “We see the demand there, and we can deploy faster.” </p><h2>GPUs shipped straight to you</h2><p>Duos Edge AI’s modular compute pods are 55 feet long and 12.5 feet wide. Though they look similar to a shipping container, they’re actually a bit larger and designed primarily for transportation by truck. Each compute pod contains racks of GPUs much like those used in other data centers. Duos recently <a href="https://ir.duostechnologies.com/news-events/press-releases/detail/830/duos-technologies-group-executes-definitive-agreement-with" target="_blank">entered</a> a deal with AI infrastructure company Hydra Host to deploy four pods with 576 GPUs per pod. That’s a total of 2,304 GPUs, with the option to later double the deployment to 4,608 GPUs. </p><p>Modular data centers aren’t new for Duos; the company previously deployed edge data centers for rural customers, <a href="https://spectrum.ieee.org/rural-data-centers" target="_self">such as the Amarillo, Texas, school district</a>. However, the pods for the Hydra Host deployment will be upgraded to handle more intense AI workloads. They’ll contain more racks, draw more power, and use liquid cooling to keep the GPUs running efficiently. <br/><br/>Across the Pacific, Korean technology giant LG is taking a similar approach. The company’s CNS subsidiary, which provides IT infrastructure and services, <a href="https://www.koreatimes.co.kr/business/tech-science/20260305/lg-cns-unveils-container-based-ai-box-for-rapid-ai-data-center-expansion">has announced the AI Modular Data Center, which</a>, like the Duos unit, contains racks of GPUs and supporting hardware in a pre-fabricated enclosure.</p><p>Also like Duos’ deployment, LG’s AI Modular Data Center contains 576 Nvidia GPUs with the option to scale up in the future. “We are currently developing an expanded version that can support more than 4,600 GPUs within a single unit, with a service launch planned within this year,” said <a href="https://www.linkedin.com/in/heonhyeock-cho-29427b147/?originalSubdomain=kr" rel="noopener noreferrer" target="_blank">Heon Hyeock Cho</a>, vice president and head of the data center business unit at LG CNS. LG’s first Modular Data Center will roll out in the South Korean port city of Busan, where it could deploy up to 50 units.</p><p>LG and Duos are not alone. <a href="https://www.hpe.com/us/en/services/ai-mod-pod.html" rel="noopener noreferrer" target="_blank">Hewlett Packard Enterprise,</a> <a href="https://www.vertiv.com/en-emea/solutions/vertiv-modular-solutions/?utm_source=press-release&utm_medium=public-relations&utm_campaign=hpc-ai&utm_content=en-coolchip" rel="noopener noreferrer" target="_blank">Vertiv</a>, and <a href="https://www.se.com/ww/en/work/solutions/data-centers-and-networks/modular-data-center/" rel="noopener noreferrer" target="_blank">Schneider Electric</a> now have modular data centers available or in development. A <a href="https://www.grandviewresearch.com/industry-analysis/modular-data-center-market-report" target="_blank">report</a> from market research firm <a href="https://www.grandviewresearch.com/" target="_blank">Grand View Research</a> estimates that the market for modular data centers could more than double by 2030.</p><h2>On the grid, but under the radar</h2><p>A modular data center site is quite different from a traditional data center because there’s no need to construct a large steel-and-concrete shell. Instead, the site can be made ready by pouring a concrete pad. The pre-fabricated modules are delivered by truck, placed on the pad where desired, and then networked on-site.<br/><br/>Duos’ deployments, for instance, include power modules placed alongside the compute pods, and the pods are networked together with redundant fiber connections that allow the pods to operate in unison. Recker compared it to lining up school buses in a parking lot. “Everything is built off-site at a factory, and we can put it together like a jigsaw puzzle,” he said.</p><p>That simplicity is the point. Both Duos and LG CNS expect a modular data center can be deployed in about six months, compared to the roughly two or three years a conventional data center requires. Recker said that, for Duos, the turnaround is so quick that building the pre-fabricated unit isn’t always the constraint. While it’s possible to construct a pre-fabricated unit in 60 or 90 days, site preparation extends the timeline “because you can’t get the permits that fast.”</p><p>Modular data centers may also provide good value. Recker said a 5-megawatt modular deployment can be built for about $25 million, and that Duos’ cost per megawatt is roughly half what larger facilities charge. For Duos, savings are possible in part because its modular data centers can target smaller deployments where the permitting is less complex. Smaller, modular deployments also meet less resistance from local governments, which are increasingly skeptical about data center construction. </p><p>While Duos targets smaller deployments, LG hopes to go big. Its planned Busan campus of 50 AI Modular Data Centers suggests an ambition to achieve deployments that rival the capacity of conventional facilities. A site with 50 units would bring the total number of GPUs to over 28,000. Here, the benefits of a modular approach could stem mostly from scalability, as a modular data center could start small and grow as required.</p><p>“By adopting a modular approach, the AI Modular Data Center can be incrementally expanded through the combination of dozens of AI Boxes,” Cho said. “It’s enabling the construction of even hyperscale-level AI data centers.”</p>]]></description><pubDate>Mon, 30 Mar 2026 14:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/modular-data-center</guid><category>Data-center</category><category>Networking</category><category>Liquid-cooling</category><category>Ai</category><dc:creator>Matthew S. Smith</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/overhead-view-of-two-data-center-pods-each-measuring-55-feet-long-by-12-5-feet-wide.jpg?id=65417343&amp;width=980"></media:content></item><item><title>Facial Recognition Is Spreading Everywhere</title><link>https://spectrum.ieee.org/facial-recognition-gone-wrong</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/illustration-34-orange-women-icons-1-blue-man-icon-labels-for-skin-tone-and-gender-comparisons.jpg?id=65407585&width=1245&height=700&coordinates=0%2C116%2C0%2C117"/><br/><br/><p>Facial recognition technology (FRT) dates back 60 years. Just over a decade ago, deep-learning methods tipped the technology into more useful—<a href="https://spectrum.ieee.org/china-facial-recognition" target="_blank">and menacing</a>—territory. Now, retailers, your neighbors, and law enforcement are all storing your face and building up a fragmentary photo album of your life.</p><p>Yet the story those photos can tell inevitably has errors. FRT makers, like those of any diagnostic technology, must balance two types of errors: false positives and false negatives. There are three possible outcomes.</p><div class="ieee-sidebar-medium"><h3>Three Possible Outcomes</h3><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="White figures and an orange hooded figure, focusing on the hooded figure in a split design." class="rm-shortcode" data-rm-shortcode-id="8a762ebf2761a791f12500ed10596cc3" data-rm-shortcode-name="rebelmouse-image" id="f4d64" loading="lazy" src="https://spectrum.ieee.org/media-library/white-figures-and-an-orange-hooded-figure-focusing-on-the-hooded-figure-in-a-split-design.png?id=65407894&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">a) identifies the suspect, since the two images are of the same person, according to the software. Success!</small></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Abstract figures: orange hoodie enlarged, white, yellow, and orange on left, black background." class="rm-shortcode" data-rm-shortcode-id="3d130b8e4c73ee49898645524cecd1f6" data-rm-shortcode-name="rebelmouse-image" id="30881" loading="lazy" src="https://spectrum.ieee.org/media-library/abstract-figures-orange-hoodie-enlarged-white-yellow-and-orange-on-left-black-background.png?id=65407867&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">b) matches another person in the footage with the suspect’s probe image. A false positive, coupled with sloppy verification, could put the wrong person behind bars and lets the real criminal escape justice.</small></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Three white icons and one orange hoodie icon on left, large orange hoodie icon on right." class="rm-shortcode" data-rm-shortcode-id="4cdaa23680c5144a5c284fcd8cb6f3df" data-rm-shortcode-name="rebelmouse-image" id="fbc8f" loading="lazy" src="https://spectrum.ieee.org/media-library/three-white-icons-and-one-orange-hoodie-icon-on-left-large-orange-hoodie-icon-on-right.png?id=65407858&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">c) fails to find a match at all. The suspect may be evading cameras, but if cameras just have low-light or bad-angle images, this creates a false negative. This type of error might let a suspect off and raise the cost of the manhunt.</small></p></div><p>In best-case scenarios—such as comparing someone’s passport photo to a photo taken by a border agent—false-negative rates are <a href="https://face.nist.gov/frte/reportcards/11/clearviewai_003.html" target="_blank">around two in 1,000 and false positives are less than one in 1 million</a>.</p><p>In the rare event you’re one of those false negatives, a border agent might ask you to show your passport and take a second look at your face. But as people ask more of the technology, more ambitious applications could lead to more catastrophic errors. Let’s say that police are searching for a suspect, and they’re comparing an image taken with a security camera with a previous “mug shot” of the suspect.</p><p>Training-data composition, differences in how sensors detect faces, and intrinsic differences between groups, such as age, all affect an algorithm’s performance. The <a href="https://assets.publishing.service.gov.uk/media/693002a4cdec734f4dff4149/1a_Cognitec_NPL_Equitability_Report_October_25.pdf&sa=D&source=docs&ust=1774557264829489&usg=AOvVaw13R0ue8NITZ-0tPVLcJ8S-" target="_blank">United Kingdom estimated</a> that its FRT exposed some groups, such as women and darker-skinned people, to risks of misidentification as high as two orders of magnitude greater than it did to others.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Five faces arranged left to right, from easy to hard to recognize." class="rm-shortcode" data-rm-shortcode-id="ce19d3eb3745de15489274ebe5083f06" data-rm-shortcode-name="rebelmouse-image" id="3ab1e" loading="lazy" src="https://spectrum.ieee.org/media-library/five-faces-arranged-left-to-right-from-easy-to-hard-to-recognize.png?id=65407777&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">Less clear photographs are harder for FRT to process.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">iStock</small></p><p>What happens with photos of people who aren’t cooperating, or vendors that train algorithms on biased datasets, or field agents who demand a swift match from a huge dataset? Here, things get murky.</p><div class="ieee-sidebar-medium"><h3>Facial Recognition Gone Wrong</h3><p><strong>THE NEGATIVES OF FALSE POSITIVES</strong></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Detroit Police SUV with American flag decal on side under bright sunlight." class="rm-shortcode" data-rm-shortcode-id="1a424f342f44dff48e8b6b05c79f5032" data-rm-shortcode-name="rebelmouse-image" id="c102c" loading="lazy" src="https://spectrum.ieee.org/media-library/detroit-police-suv-with-american-flag-decal-on-side-under-bright-sunlight.png?id=65407650&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">2020: <a href="https://quadrangle.michigan.law.umich.edu/issues/winter-2024-2025/flawed-facial-recognition-technology-leads-wrongful-arrest-and-historic&sa=D&source=docs&ust=1774557264902408&usg=AOvVaw3xUv5_o_zg1Fh0EScZ9lTW" target="_blank">Robert Williams’s wrongful arrest</a> cost him detention. The ensuing settlement requires Detroit police to enact policies that recognize FRT’s limits. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">iStock</small></p><p><strong>ALGORITHMIC BIAS</strong></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt='Red sign reads "Security cameras in use" with camera graphic.' class="rm-shortcode" data-rm-shortcode-id="014ac05f2fe587ca01643c64c750e331" data-rm-shortcode-name="rebelmouse-image" id="f4f1f" loading="lazy" src="https://spectrum.ieee.org/media-library/red-sign-reads-security-cameras-in-use-with-camera-graphic.png?id=65407620&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">2023: <a href="https://incidentdatabase.ai/cite/619/&sa=D&source=docs&ust=1774557264903427&usg=AOvVaw3fBw_78OyUB3Sa_cPpxmCi" target="_blank">Court bans Rite Aid from using facial recognition for five years</a> over its use of a racially biased algorithm. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">iStock</small></p><p><strong>TOO FAST, TOO FURIOUS?</strong></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Back of ICE officer in tactical gear facing a house." class="rm-shortcode" data-rm-shortcode-id="0004b023a075c21698cdf88cfd0b4106" data-rm-shortcode-name="rebelmouse-image" id="889f9" loading="lazy" src="https://spectrum.ieee.org/media-library/back-of-ice-officer-in-tactical-gear-facing-a-house.png?id=65407619&width=980"/><small class="image-media media-caption" placeholder="Add Photo Caption...">2026: U.S. immigration agents <a href="https://www.404media.co/ices-facial-recognition-app-misidentified-a-woman-twice/&sa=D&source=docs&ust=1774557264904407&usg=AOvVaw03DUrBl3YxN6c3uhHa611f" target="_blank">misidentify a woman they’d detained as two different women</a>. </small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">VICTOR J. BLUE/BLOOMBERG/GETTY IMAGES </small></p></div><p><span>Consider a busy trade fair using FRT to check attendees against a database, or gallery, of images of the 10,000 registrants, for example. Even at 99.9 percent accuracy you’ll get about a dozen false positives or negatives, which may be worth the trade-off to the fair organizers. But if police start using something like that across a city of 1 million people, the number of potential victims of mistaken identity rises, as do the stakes.</span></p><p><span>What if we ask FRT to tell us if the government has ever recorded and stored an image of a given person? That’s what U.S. Immigration and Customs Enforcement <a href="https://illinoisattorneygeneral.gov/News-Room/Current-News/001%20-%20Complaint%201.12.26.pdf?language_id=1" target="_blank">agents have done since June 2025</a>, using the Mobile Fortify app. The agency conducted more than 100,000 FRT searches in the first six months. The size of the potential gallery is at least <a href="https://sam.gov/opp/b016354c5bd045fa92e4886878747dc8/view" target="_blank">1.2 billion images</a>.</span></p><p><span>At that size, assuming even best-case images, the system is likely to return around 1 million false matches, but at a rate at least 10 times as high for darker-skinned people, depending on the subgroup.</span></p><p>Responsible use of this powerful technology would involve independent identity checks, multiple sources of data, and a clear understanding of the error thresholds, says computer scientist <a href="https://www.cics.umass.edu/about/directory/erik-learned-miller" target="_blank">Erik Learned-Miller</a> of the University of Massachusetts Amherst: “<a href="https://spectrum.ieee.org/joy-buolamwini" target="_blank">The care we take</a> in deploying such systems should be proportional to the stakes.”</p>]]></description><pubDate>Mon, 30 Mar 2026 13:00:02 +0000</pubDate><guid>https://spectrum.ieee.org/facial-recognition-gone-wrong</guid><category>Facial-recognition</category><category>Privacy</category><category>Surveillance</category><category>Machine-vision</category><category>Computer-vision</category><dc:creator>Lucas Laursen</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/illustration-34-orange-women-icons-1-blue-man-icon-labels-for-skin-tone-and-gender-comparisons.jpg?id=65407585&amp;width=980"></media:content></item><item><title>NYU’s Quantum Institute Bridges Science and Application</title><link>https://spectrum.ieee.org/nyu-quantum-institute</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/person-in-white-suit-working-with-semiconductor-equipment-in-a-lab.jpg?id=65322091&width=1245&height=700&coordinates=0%2C0%2C0%2C0"/><br/><br/><p><em>This sponsored article is brought to you by <a href="https://engineering.nyu.edu/" rel="noopener noreferrer" target="_blank">NYU Tandon School of Engineering</a>.</em></p><p>Within a 6 mile radius of New York University’s (NYU) campus, there are more than 500 tech industry giants, banks, and hospitals. This isn’t just a fact about real estate, it’s the foundation for advancing quantum discovery and application.</p><p>While the world races to harness quantum technology, NYU is betting that the ultimate advantage lies not solely in a lab, but in the dense, demanding, and hyper-connected urban ecosystem that surrounds it. With the launch of its <a href="https://www.nyu.edu/about/news-publications/news/2025/october/nyu-launches-quantum-institute-.html" rel="noopener noreferrer" target="_blank"><span>NYU Quantum Institute</span></a> (NYUQI), NYU is positioning itself as <a href="https://www.nyu.edu/about/news-publications/news/2025/october/top-quantum-scientists-convene-at-nyu.html" target="_blank">the central node</a> in this network; a “full stack” powerhouse built on the conviction that it has found the right place, and the right time, to turn quantum science into tangible reality.</p><p>Proximity advantage is essential because quantum science demands it. Globally, the quest for practical quantum solutions — whether for computing, sensing, or secure communications — has been stalled, in part, by fragmentation. Physicists and chemical engineers invent new materials, computer scientists develop new algorithms, and electrical engineers build new devices, but all three often work in isolated academic silos.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Three men pose at the 4th Annual NYC Quantum Summit 2025; attendees converse in the background." class="rm-shortcode" data-rm-shortcode-id="1dd6dfe45b73630bb9040545fcdfae7d" data-rm-shortcode-name="rebelmouse-image" id="33e2d" loading="lazy" src="https://spectrum.ieee.org/media-library/three-men-pose-at-the-4th-annual-nyc-quantum-summit-2025-attendees-converse-in-the-background.jpg?id=65322345&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Gregory Gabadadze, NYU’s dean for science, NYU physicist and Quantum Institute Director Javad Shabani, and Juan de Pablo, Anne and Joel Ehrenkranz Executive Vice President for Global Science and Technology and executive dean of the Tandon School of Engineering.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Veselin Cuparić/NYU</small></p><p><span>NYUQI’s premise is that breakthroughs happen “at the interfaces between different domains,” according to </span><a href="https://engineering.nyu.edu/faculty/juan-de-pablo" target="_blank"><span>Juan de Pablo</span></a><span>, Executive Vice President for Global Science and Technology at NYU and Executive Dean of the NYU Tandon School of Engineering. The Institute is built to actively force those necessary collisions — to integrate the physicists, engineers, materials scientists, computer scientists, biologists, and chemists vital to quantum research into one holistic operation. This institutional design ensures that the hardware built by one team can be immediately tested by software developed by another, accelerating progress in a way that isolated departments never could.</span></p><p class="pull-quote"><span>NYUQI’s premise is that breakthroughs happen at the interfaces between different domains. <strong>—Juan de Pablo, NYU Tandon School of Engineering</strong></span></p><p>NYUQI’s integrated vision is backed by a massive physical commitment to the city. The NYUQI is not just a theoretical concept; its collaborators will be housed in a renovated, <a href="https://www.nyu.edu/about/news-publications/news/2025/may/nyu-entering-long-term-lease-at-770-broadway.html" target="_blank"><span>million-square-foot facility</span></a> in the heart of Manhattan’s West Village, backed by a state-of-the-art <a href="https://engineering.nyu.edu/research/nanofab" target="_blank">Nanofabrication Cleanroom</a> in Brooklyn serving as a high-tech foundry. This is where the theoretical meets physical devices, allowing the Institute to test and refine the process from materials science to deployment.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt='NYU building exterior with "Science + Tech" signage, flags, and a passing yellow taxi.' class="rm-shortcode" data-rm-shortcode-id="605cc71d844927d3fb0a05fb086fedcf" data-rm-shortcode-name="rebelmouse-image" id="bceaa" loading="lazy" src="https://spectrum.ieee.org/media-library/nyu-building-exterior-with-science-tech-signage-flags-and-a-passing-yellow-taxi.jpg?id=65322352&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">NYUQI will be housed in a renovated, million-square-foot facility in the heart of Manhattan’s West Village.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Tracey Friedman/NYU</small></p><p><span>Leading this effort is NYUQI Director </span><a href="https://as.nyu.edu/faculty/javad-shabani.html" target="_blank"><span>Javad Shabani</span></a><span>, who, along with the other members, is turning the Institute into a hub for collaboration with private and public sector partners with quantum challenges that need solving. As de Pablo explains, “Anybody who wants to work on quantum with NYU, you come in through that door, and we’ll send you to the right place.” For New York’s vast ecosystem of tech giants and financial institutions, the NYUQI offers a resource they can’t build on their own: a cohesive team of experts in quantum phenomena, quantum information theory, communication, computing, materials, and optics, and a structured path to applying theoretical discoveries to advanced quantum technologies.</span></p><h2>Solving the Challenge of Quantum Research</h2><p><span>The NYUQI’s integrated structure is less about organizational management, and more about scientific requirement. </span><span>The challenge of quantum is that the hardware, the software, and the programming are inherently interconnected — each must be designed to work with the other. To solve this, the Institute focuses on three applications of quantum science: Quantum Computing, Quantum Sensing, and Quantum Communications.</span></p><p>For Shabani, this means creating an integrated environment that bridges discovery with experimentation, starting with the physical components all the way to quantum algorithm centers. That will include a fabrication facility in the new building in Manhattan, as well as the <a href="https://engineering.nyu.edu/news/chips-and-science-act-spurs-nanofab-cleanroom-ribbon-cutting-nyu-tandon-school-engineering" target="_blank"><span>NYU Nanofab</span></a> in Brooklyn directed by Davood Shahjerdi. New York Senators Charles Schumer and Kirsten Gillibrand recently secured <a href="https://www.nyu.edu/about/news-publications/news/2026/february/nyu-receives--1-million-in-funding-from-senators-schumer-and-gil.html" target="_blank">$1 million in congressionally-directed spending</a> to bring Thermal Laser Epitaxy (TLE) technology — which allows for atomic-level purity, minimal defects, and streamlined application of a diverse range of quantum materials — to NYU, marking the first time the equipment will be used in the U.S.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Two people hold semiconductor wafers during a presentation with audience taking photos." class="rm-shortcode" data-rm-shortcode-id="1a0dbca6c6bb8fb7dbf4d399689b2922" data-rm-shortcode-name="rebelmouse-image" id="d434c" loading="lazy" src="https://spectrum.ieee.org/media-library/two-people-hold-semiconductor-wafers-during-a-presentation-with-audience-taking-photos.jpg?id=65322354&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">NYU Nanofab manager Smiti Bhattacharya and Nanofab Director Davood Shahjerdi at the nanofab ribbon-cutting in 2023. The nanofab is the first academic cleanroom in Brooklyn, and serves as a prototyping facility for the NORDTECH Microelectronics Commons consortium.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">NYU WIRELESS</small></p><p>Tight control over fabrication, and can allow researchers to pivot quickly when a breakthrough in one area — say, finding a cheaper, more reliable material like silicon carbide — can be explored for use across all three applications, and offers unique access to academics and the private sector alike to sophisticated pieces of specialty equipment whose maintenance knowledge and costs make them all-but-impossible to maintain outside of the right staffing and environment.</p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="3D model of a laboratory layout, highlighting the Yellow Room in bright yellow." class="rm-shortcode" data-rm-shortcode-id="e7c1128703d96de919ed2ce440a97416" data-rm-shortcode-name="rebelmouse-image" id="62d58" loading="lazy" src="https://spectrum.ieee.org/media-library/3d-model-of-a-laboratory-layout-highlighting-the-yellow-room-in-bright-yellow.png?id=65322596&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">The NYU Nanofab is Brooklyn’s first academic cleanroom, with a strategic focus on superconducting quantum technologies, advanced semiconductor electronics, and devices built from quantum heterostructures and other next-generation materials.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">NYU Nanofab</small></p><p><span>That speed and adaptability is the NYUQI’s competitive edge. It turns fragmented challenges into holistic solutions, positioning the Institute to solve real-world problems for its New York neighbors—from highly secure data transmission to next-generation drug discovery.</span></p><h2>Testing Quantum Communication in NYC</h2><p>The integrated approach also makes the NYUQI a testbed for the most critical near-term applications. Take Quantum Communications, which is essential for creating an “unhackable” quantum internet. In an industry first, NYU worked with the quantum start-up Qunnect to <a href="https://www.nyu.edu/about/news-publications/news/2023/september/nyu-takes-quantum-step-in-establishing-cutting-edge-tech-hub-in-.html" target="_blank"><span>send quantum information through standard telecom fiber</span></a> in New York City between Manhattan and Brooklyn through a 10-mile quantum networking link. Instead of simulating communication challenges in a lab, the NYUQI team is already leveraging NYU’s city-wide campus by utilizing existing infrastructure to test secure quantum transmission between Manhattan and Brooklyn. </p><p class="pull-quote">The NYUQI team is already leveraging NYU’s city-wide campus by utilizing existing infrastructure to test secure quantum transmission between Manhattan and Brooklyn.</p><p>This isn’t just theory; it is building a functioning prototype in the most demanding, dense urban environment  in the world. Real-time, real-world deployment is a critical component missing in other isolated institutions. When the NYUQI achieves results, the technology will be that much more readily available to the massive financial, tech, and communications organizations operating right outside their door.</p><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" style="float: left;"> <img alt="Scientist in protective gear working in a laboratory with samples." class="rm-shortcode" data-rm-shortcode-id="d644b791788af64769a853d0516834e6" data-rm-shortcode-name="rebelmouse-image" id="dc2fb" loading="lazy" src="https://spectrum.ieee.org/media-library/scientist-in-protective-gear-working-in-a-laboratory-with-samples.jpg?id=65322378&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">NYUQI includes a state-of-the-art Nanofabrication Cleanroom in Brooklyn serving as a high-tech foundry.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">NYU Tandon</small></p><p><span>While the Institute has built the physical infrastructure and designed the necessary scientific architecture, its enduring contribution will be the specialized workforce it creates for the new quantum economy. This addresses the market’s greatest deficit: a lack of individuals trained not just in physics, but in the integrated, full-stack approach that quantum demands.</span></p><p>By creating a pipeline of 100 to 200 graduate and doctoral students who are encouraged to collaborate across Computing, Sensing, and Communications, the NYUQI is narrowing the skills gap. These will be future leaders who can speak the language of the physicist, the materials scientist, and the engineer simultaneously. This commitment to interdisciplinary talent is also fueled by the launch of the new Master of Science in Quantum Science & Technology program at NYU Tandon, positioning the university among a select group worldwide offering such a specialized degree.</p><p>Interdisciplinary education creates the shared language and understanding poised to make graduates coming from collaborations in the NYUQI extremely valuable in the current landscape. Quantum challenges are not just technical; they are managerial and philosophical as well. An engineer working with the NYUQI will understand the requirements of the nanofabrication cleanroom and the foundations of superconducting qubits for quantum computing, just as a physicist will understand the application needs of an industry partner like a large financial institution. In a field where the entire team must be able to communicate seamlessly, these are professionals truly equipped to rapidly translate discovery into deployable technology. Creating a talent pipeline at scale will provide a missing link that converts New York’s vast commercial energy into genuine quantum advantage.</p><h2>NYUQI: Building Talent, Technology, and Structure</h2><p><span>The vision for the NYUQI </span><span>is an act of strategic geography that plays directly into the sheer volume of opportunity and demand right outside their new facility. </span><span>By building the talent, the technology, and the structure necessary to capitalize on this dense environment, NYU is not just participating in the quantum race, it is actively steering it.</span></p><p class="shortcode-media shortcode-media-rebelmouse-image"> <img alt="Conference room with attendees seated at round tables, facing a presenter on stage." class="rm-shortcode" data-rm-shortcode-id="f5e2ae16e0c5ebc4f0828d52ed639115" data-rm-shortcode-name="rebelmouse-image" id="02b7e" loading="lazy" src="https://spectrum.ieee.org/media-library/conference-room-with-attendees-seated-at-round-tables-facing-a-presenter-on-stage.jpg?id=65322370&width=980"/> <small class="image-media media-caption" placeholder="Add Photo Caption...">Attendees of NYU’s 2025 Quantum Summit.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">Tracey Friedman/NYU</small></p><p>The initial hypothesis for the NYUQI was simple: the ultimate advantage lies in pursuing the science in the right place at the right time. Now, the institute will ensure that the next wave of scientific discovery, capable of solving previously intractable problems in finance, medicine, and security, will be conceived, built, and tested in the heart of New York City.</p>]]></description><pubDate>Fri, 27 Mar 2026 10:02:05 +0000</pubDate><guid>https://spectrum.ieee.org/nyu-quantum-institute</guid><category>Nyu-tandon</category><category>Quantum-computing</category><category>Quantum-internet</category><category>Semiconductors</category><category>Quantum-communications</category><dc:creator>Wiley</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/person-in-white-suit-working-with-semiconductor-equipment-in-a-lab.jpg?id=65322091&amp;width=980"></media:content></item><item><title>IEEE 802.11bn Delivers Ultra-High Reliability for Wi-Fi 8</title><link>https://content.knowledgehub.wiley.com/setting-new-performance-standards-with-ieee-802-11bn-an-in-depth-overview-of-wi-fi-8/</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/logo-of-rohde-schwarz-with-slogan-make-ideas-real-and-stylized-rs-in-a-diamond-shape.png?id=65355284&width=980"/><br/><br/><p><span>A technical exploration of IEEE 802.11bn’s physical and MAC layer enhancements — including distributed resource units, enhanced long range, multi-AP coordination, and seamless roaming — that define Wi-Fi 8.</span></p><p><strong><span>What Attendees will Learn</span></strong></p><ol><li><span>Why Wi-Fi 8 prioritizes reliability over raw throughput — Understand how IEEE 802.11bn shifts the design philosophy from peak data-rate gains to ultra-high reliability.</span></li><li>How new physical layer features overcome uplink power limitations — Learn how distributed resource units spread tones across wider distribution bandwidths to boost per-tone transmit power, and how enhanced long range protocol data units use power-boosted preamble fields and frequency-domain duplication to extend uplink coverage.</li><li>How advanced MAC coordination reduces interference and latency — Examine multi-access point coordination schemes — coordinated beamforming, spatial reuse, time division multiple access, and restricted target wake time — alongside non-primary channel access and priority enhanced distributed channel access.</li><li>What seamless roaming and power management mean for next-generation deployments — Discover how seamless mobility domains eliminate reassociation delays during access point transitions, and how dynamic power save and multi-link power management let devices trade capability for battery life without sacrificing connectivity.</li></ol><p><a href="https://content.knowledgehub.wiley.com/setting-new-performance-standards-with-ieee-802-11bn-an-in-depth-overview-of-wi-fi-8/" target="_blank">Download this free whitepaper now!</a></p>]]></description><pubDate>Wed, 25 Mar 2026 14:22:07 +0000</pubDate><guid>https://content.knowledgehub.wiley.com/setting-new-performance-standards-with-ieee-802-11bn-an-in-depth-overview-of-wi-fi-8/</guid><category>Wifi</category><category>Internet</category><category>Standards</category><category>Transmission</category><category>Type-whitepaper</category><dc:creator>Rohde &amp; Schwarz</dc:creator><media:content medium="image" type="image/png" url="https://assets.rbl.ms/65355284/origin.png"></media:content></item><item><title>Data Centers Are Transitioning From AC to DC</title><link>https://spectrum.ieee.org/data-center-dc</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/nvidia-s-high-compute-density-racks.jpg?id=65397940&width=1245&height=700&coordinates=0%2C469%2C0%2C469"/><br/><br/><p>Last week’s <a href="https://www.nvidia.com/gtc/" target="_blank">Nvidia GTC</a> conference highlighted new <a href="https://spectrum.ieee.org/nvidia-groq-3" target="_blank">chip</a> architectures to power AI. But as the chips become faster and more powerful, the remainder of data center <a data-linked-post="2674166715" href="https://spectrum.ieee.org/data-center-liquid-cooling" target="_blank">infrastructure</a> is playing catch-up. The power-delivery community  is responding: Announcements from <a href="https://www.prnewswire.com/news-releases/delta-exhibits-energy-saving-solutions-for-800-vdc-in-next-gen-ai-factories-and-digital-twin-applications-built-on-omniverse-at-nvidia-gtc-2026-302715850.html" rel="noopener noreferrer" target="_blank">Delta</a>,  <a href="https://www.eaton.com/us/en-us/company/news-insights/news-releases/2026/eaton-collaborates-with-nvidia-to-unveil-its-beam-rubin-dsx-platform.html" rel="noopener noreferrer" target="_blank">Eaton</a>, <a href="https://www.se.com/us/en/about-us/newsroom/news/press-releases/Schneider-Electric-teams-with-NVIDIA-to-develop-validated-blueprints-to-design-simulate-build-operate-and-maintain-gigawattscale-AI-Factories-69b82f61aa1027e04205d273/" target="_blank">Schneider Electric</a>, and <a href="https://www.vertiv.com/en-us/about/news-and-insights/corporate-news/2026/vertiv-brings-converged-physical-infrastructure-to-nvidia-vera-rubin-dsx-ai-factories/" rel="noopener noreferrer" target="_blank">Vertiv</a> showcased new designs for the AI era. Complex and inefficient AC-to-DC power conversions are gradually being replaced by DC configurations, at least in hyperscale data centers.</p><p>“While AC distribution remains deeply entrenched, advances in power electronics and the rising demands of AI infrastructure are accelerating interest in DC architectures,” says <a href="https://www.linkedin.com/in/solarchris/" target="_blank">Chris Thompson</a>, vice president of advanced technology and global microgrids at Vertiv.</p><h2>AC-to-DC Conversion Challenges</h2><p>Today, nearly all data centers are designed around AC utility power. The electrical path includes multiple conversions before power reaches the compute load. Power typically enters the data center as medium-voltage AC (1 to 35 kilovolts), is stepped down to low-voltage AC (480 or 415 volts) using a transformer, converted to DC inside an uninterruptible power supply (UPS) for battery storage, converted back to AC, and converted again to low-voltage DC (typically 54 V DC) at the server, supplying the DC power computing chips actually require.</p><p>“The double conversion process ensures the output AC is clean, stable, and suitable for data center servers,” says <a href="https://www.linkedin.com/in/luiz-fernando-huet-de-bacellar-b2112117/" target="_blank">Luiz Fernando Huet de Bacellar,</a> vice president of engineering and technology at Eaton.</p><p>That setup worked well enough for the amounts of power required by traditional data centers. Traditional data center computational racks draw on the order of 10 kW each. For AI, that is starting to approach 1 megawatt.  At that scale, the energy losses, current levels, and copper requirements of AC-to-DC conversions become increasingly difficult to justify. Every conversion incurs some power loss. On top of that, as the amount of power that needs to be delivered grows, the sheer size of the convertors, as well as the connector requirements of copper busbars, becomes untenable.<span> According to an Nvidia <a href="https://developer.nvidia.com/blog/nvidia-800-v-hvdc-architecture-will-power-the-next-generation-of-ai-factories/" target="_blank">blog</a>, a 1-MW rack</span><span> could require as much as 200 kilograms of copper busbar. For a 1-gigawatt data center, it could amount to 200,000 kg of copper. </span></p><h2>Benefits of High-Voltage DC Power</h2><p>By converting 13.8-kV AC grid power directly to 800 V DC at the data center perimeter, most intermediate conversion steps are eliminated. This reduces the number of fans and power-supply units, and leads to higher system reliability, lower heat dissipation, improved energy efficiency, and a smaller equipment footprint.</p><p>“Each power conversion between the electric grid or power source and the silicon chips inside the servers causes some energy loss,” says Bacellar.</p><p>Switching from 415-V AC to 800-V DC in electrical distribution enables 85 percent more power to be transmitted through the same conductor size. This happens because higher voltage reduces current demand, lowering resistive losses and making power transfer more efficient. Thinner conductors can handle the same load, reducing copper requirements by 45 percent, a 5 percent improvement in efficiency, and 30 percent lower total cost of ownership for gigawatt-scale facilities.</p><p>“In a high-voltage DC architecture, power from the grid is converted from medium-voltage AC to roughly 800-V DC and then distributed throughout the facility on a DC bus,” said Vertiv’s Thompson. “At the rack, compact DC-to-DC converters step that voltage down for GPUs and CPUs.”</p><p>A <a href="https://www.datacenter-asia.com/wp-content/uploads/2025/08/Omdia-Analysts-Summit-Omdia%E5%88%86%E6%9E%90%E5%B8%88%E5%B3%B0%E4%BC%9A.pdf" target="_blank">report</a> from technology advisory group <a href="https://omdia.tech.informa.com/" target="_blank">Omdia</a> claims that higher voltage DC data centers have already appeared in China. In the Americas, the <a href="https://www.linkedin.com/posts/sharada-yeluri_microsoft-meta-google-activity-7367974455052017666-nXV5/" target="_blank">Mt. Diablo Initiative</a> (a collaboration among <a href="https://www.meta.com/about/?srsltid=AfmBOoq7uBjCU2oG3oI6Ti8VQaMdaxhAcxXmXD-twy9OTi0cbmTqGKVQ" target="_blank">Meta</a>, <a href="https://www.microsoft.com/en-us" target="_blank">Microsoft</a>, and the <a href="https://www.opencompute.org/" target="_blank">Open Compute Project</a>) is a 400-V DC rack power distribution experiment.</p><h2>Innovations in DC Power Systems</h2><p>A handful of vendors are trying to get ahead of the game. Vertiv’s 800-V DC ecosystem that integrates with <a href="https://www.vertiv.com/en-us/about/news-and-insights/corporate-news/vertiv-develops-energy-efficient-cooling-and-power-reference-architecture-for-the-nvidia-gb300-nvl72/" target="_blank">Nvidia Vera Rubin Ultra Kyber platforms</a> will be commercially available in the second half of 2026. Eaton, too, is well advanced in its 800-V DC systems innovation courtesy of a medium-voltage solid-state transformer (SST) that will sit at the heart of DC power distribution system. Meanwhile Delta, has released 800-V DC in-row 660-kW power racks with a total of 480 kW of embedded battery backup units. And, <a href="https://www.solaredge.com/us/" target="_blank">SolarEdge</a> is hard at work on a 99%-efficient SST that will be paired with a native DC UPS and a DC power distribution layer.</p><p>But much of the industry is far behind. <a href="https://www.linkedin.com/in/pehughes/" target="_blank">Patrick Hughes</a>, senior vice president of strategy, technical, and industry affairs for the <a href="https://www.makeitelectric.org/" target="_blank">National Electrical Manufacturers Association</a>, says most innovation is happening at the 400-V DC level, though some are preparing 800-V DC. He believes the industry needs a complete, coordinated ecosystem, including power electronics, protection, connectors, sensing, and service‑safe components that scale together rather than in isolation. That, in turn, requires retooling manufacturing capacity for DC‑specific equipment, expanding semiconductor and materials supply, and clear, long‑term demand commitments that justify major capital investment across the value chain.</p><p>“Many are taking a cautious approach, offering limited or adapted solutions while waiting for clearer standards, safety frameworks, and customer commitments,” said Hughes. “Building the supply chain will hinge on stabilizing standards and safety frameworks so suppliers can design, certify, manufacture, and install equipment with confidence.”</p>]]></description><pubDate>Tue, 24 Mar 2026 16:00:05 +0000</pubDate><guid>https://spectrum.ieee.org/data-center-dc</guid><category>Data-centers</category><category>Power-electronics</category><category>Ai</category><dc:creator>Drew Robb</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/nvidia-s-high-compute-density-racks.jpg?id=65397940&amp;width=980"></media:content></item><item><title>What Will It Take to Build the World’s Largest Data Center?</title><link>https://spectrum.ieee.org/5gw-data-center</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/construction-symbols-on-yellow-background.png?id=65356154&width=1245&height=700&coordinates=0%2C973%2C0%2C974"/><br/><br/><p><strong>The undying thirst for </strong>smarter (historically, that means larger) AI models and greater adoption of the ones we already have has led to an explosion in <a href="https://epoch.ai/data/data-centers#data-insights" rel="noopener noreferrer" target="_blank">data-center construction projects</a>, unparalleled both in number and scale. Chief among them is Meta’s planned 5-gigawatt data center in Louisiana, called Hyperion, announced in June of 2025. Meta CEO Mark Zuckerberg said Hyperion will “cover a significant part of the footprint of Manhattan,” and the first phase—a 2-GW version—will be completed by 2030.</p><p>Though the project’s stated 5-GW scale is the largest among its peers, it’s just one of several dozen similar projects now underway. According to Michael Guckes, chief economist at construction-software company <a href="https://www.constructconnect.com/preconstruction-software?campaign=21011210878&group=161161401080&target=kwd-337013613104&matchtype=e&creative=760058507701&device=c&se_kw=constructconnect&utm_medium=ppc&utm_campaign=CC+Brand+2&utm_term=constructconnect&utm_source=adwords&hsa_ad=760058507701&hsa_kw=constructconnect&hsa_net=adwords&hsa_tgt=kwd-337013613104&hsa_grp=161161401080&hsa_src=g&hsa_ver=3&hsa_cam=21011210878&hsa_mt=e&hsa_acc=3324869874&gad_source=1&gad_campaignid=21011210878&gbraid=0AAAAADccs_biRlt8tR8-qu3h7Kja1Tzte&gclid=CjwKCAiA3-3KBhBiEiwA2x7FdCQc4sQOa0YZVFnCW9RF1tGkH2hDiowNrjM587XsXAv6Fb7Sdr1hgBoCNjEQAvD_BwE" rel="noopener noreferrer" target="_blank">ConstructConnect</a>, spending on data centers topped US $27 billion by July of 2025 and, once the full-year figures are tallied, will easily exceed $60 billion. Hyperion alone accounts for about a quarter of that.</p><p>For the engineers assigned to bring these projects to life, the mix of challenges involved represent a unique moment. The world’s largest tech companies are opening their wallets to pay for new innovations in compute, cooling, and <a data-linked-post="2674861846" href="https://spectrum.ieee.org/nvidia-rubin-networking" target="_blank">network</a> technology designed to operate at a scale that would’ve seemed absurd five years ago.</p><p>At the same time, the breakneck pace of building comes paired with serious problems. Modern data-center construction frequently requires an influx of temporary workers and sharply increases noise, traffic, pollution, and often local electricity prices. And the environmental toll remains a concern long after facilities are built due to the unprecedented 24/7 energy demands of AI data centers which, according to one recent study, <a href="https://www.nature.com/articles/s41893-025-01681-y" rel="noopener noreferrer" target="_blank">could emit the equivalent of tens of millions of tonnes of CO<span><sub>2</sub></span> annually</a> in the United States alone.</p><p>Regardless of these issues, large AI companies, and the engineers they hire, are going full steam ahead on giant data-center construction. So, what does it really take to build an unprecedentedly large data center?</p><h2>AI Rewrites Building Design</h2><p>The stereotypical data-center building rests on a reinforced concrete slab foundation. That’s paired with a steel skeleton and poured concrete wall panels. The finished building is called a “shell,” a term that implies the structure itself is a secondary concern. Meta has <a href="https://www.datacenterdynamics.com/en/news/meta-brings-data-centers-in-tents-to-gallatin-tennessee/" target="_blank">even used gigantic tents</a> to throw up temporary data centers.</p><p>Still, the scale of the largest AI data centers brings unique challenges. “The biggest challenge is often what’s under the surface. Unstable, corrosive, or expansive soils can lead to delays and require serious intervention,” says <a href="https://www.jacobs.com/our-people/meet-bob-haley" target="_blank">Robert Haley</a>, vice president at construction consulting firm <a href="https://www.jacobs.com/" target="_blank">Jacobs</a>.<a href="https://www.stantec.com/en/people/c/carter-amanda" target="_blank"> Amanda Carter</a>, a senior technical lead at <a href="https://www.stantec.com/en" target="_blank">Stantec</a>, said a soil’s thermal conductivity is also important, as most electrical infrastructure is placed underground. “If the soil has high thermal resistivity, it’s going to be difficult to dissipate [heat].” Engineers may take hundreds or thousands of soil samples before construction can begin.</p><h3>GPUs</h3><br/><img alt="Yellow microchip icon on a black background." class="rm-shortcode" data-rm-shortcode-id="9612db5baec52cce6fe11d703e52c7bc" data-rm-shortcode-name="rebelmouse-image" id="af54d" loading="lazy" src="https://spectrum.ieee.org/media-library/yellow-microchip-icon-on-a-black-background.png?id=65347639&width=980"/><p>Modern AI data centers often use <em><em>rack-scale</em></em> systems, such as the Nvidia GB200 NVL72, which occupy a single data-center rack. Each rack contains 72 GPUs, 36 CPUs, and up to 13.4 terabytes of GPU memory. The racks measure over 2.2 meters tall and weigh over one and a half tonnes, forcing AI data centers to use thicker concrete with more reinforcement to bear the load.</p><p>A single GB200 rack can use up to 120 kilowatts. If Hyperion meets its 5-gigawatt goals, the data-center campus could include over 41,000 rack-scale systems, for a total of more than 3 million GPUs. The final number of GPUs used by Hyperion is likely to be less than that, though only because future GPUs will be larger, more capable, and use more power.</p><h3>Money</h3><br/><img alt="Black hand and dollar symbol combined on an orange background." class="rm-shortcode" data-rm-shortcode-id="2ef34f3679a3b3135244243e46ae5630" data-rm-shortcode-name="rebelmouse-image" id="248eb" loading="lazy" src="https://spectrum.ieee.org/media-library/black-hand-and-dollar-symbol-combined-on-an-orange-background.png?id=65347751&width=980"/><p>According to ConstructConnect, spending on data centers neared US $27 billion through July of 2025 and, according to the latest data, will tally close to $60 billion through the end of the year. Meta’s Hyperion project is a big slice of the pie, at $10 billion.</p><p>Data-center spending has become an important prop for the construction industry, which is seeing reduced demand in other areas, such as residential construction and public infrastructure. ConstructConnect’s third quarter 2025 financial report stated that the quarter’s decline “would have been far more severe without an $11 billion surge in data center starts.”</p><h3></h3><br/><p>There’s apparently no shortage of eligible sites, however, as both the number of data centers under construction, and the money spent on them, has skyrocketed. The spending has allowed companies building data centers to throw out the rule book. Prior to the AI boom, most data centers relied on tried-and-true designs that prioritized inexpensive and efficient construction. Big tech’s willingness to spend has shifted the focus to speed and scale.</p><p>The loose purse strings open the door to larger and more robust prefabricated concrete wall and floor panels. <a href="https://www.linkedin.com/in/dougbevier/" target="_blank">Doug Bevier</a>, director of development at <a href="https://www.clarkpacific.com/" rel="noopener noreferrer" target="_blank">Clark Pacific</a>, says some concrete floor panels may now span up to 23 meters and need to handle floor loads up to 3,000 kilograms per square meter, <a href="https://codes.iccsafe.org/s/IBC2018/chapter-16-structural-design/IBC2018-Ch16-Sec1607.1" rel="noopener noreferrer" target="_blank">which is more than twice the load international building codes normally define for manufacturing and industry</a>. In some cases, the concrete panels must be custom-made for a project, an expensive step that the economics of pre-AI data centers rarely justified.</p><p>Simultaneously, the time scale for projects is also compressed: <a href="https://www.linkedin.com/in/jamiemcgrath365/" rel="noopener noreferrer" target="_blank">Jamie McGrath</a>, senior vice president of data-center operations at<a href="https://www.crusoe.ai/" rel="noopener noreferrer" target="_blank"> Crusoe</a>, says the company is delivering projects in “about 12 months,” compared to 30 to 36 months before. Not all projects are proceeding at that pace, but speed is universally a priority.</p><p>That makes it difficult to coordinate the labor and materials required. Meta’s Hyperion site, located in rural Richland Parish, Louisiana, is emblematic of this challenge. <a href="https://www.nola.com/news/business/meta-louisiana-ai-data-center/article_77f553ff-c272-4e6c-a775-60bbbee0b065.html" rel="noopener noreferrer" target="_blank">As reported by NOLA.com</a>, at least 5,000 temporary workers have flocked to the area, which has only about 20,000 permanent residents. These <a href="https://www.wsj.com/business/data-centers-are-a-gold-rush-for-construction-workers-6e3c5ce0?st=jr1y94" rel="noopener noreferrer" target="_blank">workers earn above-average wages</a> and bring a short-term boost for some local businesses, such as restaurants and convenience stores. However, they have also spurred complaints from residents about traffic and construction noise and pollution.</p><p>This friction with residents includes not only these obvious impacts, but <a href="https://youtu.be/DGjj7wDYaiI?si=aZocXHJe0IYUkJcl&t=175" rel="noopener noreferrer" target="_blank">also things you might not immediately suspect</a>, such as light pollution caused by around-the-clock schedules. Also significant are changes to local water tables and runoff, which can reduce water quality for neighbors who rely on well water. These issues have motivated a few U.S. cities <a href="https://www.atlantanewsfirst.com/2025/06/04/atlanta-tightens-restrictions-data-centers-bans-them-some-neighborhoods/" rel="noopener noreferrer" target="_blank">to enact data-center bans</a>.</p><h2>Data Centers Often Go BYOP (bring your own power)</h2><p>Meta’s Richland Parish site also highlights a problem that’s priority No. 1 for both AI data centers and their critics: power.</p><p>Data centers have always drawn large amounts of power, which nudged data-center construction to cluster in hubs where local utilities were responsive to their demands. Virginia’s electric utility, Dominion Energy, met demand with agreements to build new infrastructure, <a href="https://rmi.org/amazon-dominion-virginia-power-reach-breakthrough-renewable-energy-agreement/" rel="noopener noreferrer" target="_blank">often with a focus on renewable energy</a>.</p><p>The power demands of the largest AI data centers, though, have caught even the most responsive utilities off guard. A report from the Lawrence Berkeley National Laboratory, in California, estimated the entire U.S. data-center industry <a href="https://eta-publications.lbl.gov/sites/default/files/lbnl-1005775_v2.pdf" rel="noopener noreferrer" target="_blank">consumed an average load of roughly 8 GW of power in 2014</a>. Today, the largest AI data-center campuses are built to handle up to a gigawatt each, and Meta’s Hyperion is projected to require 5 GW.</p><p>“Data centers are exasperating issues for a lot of utilities,” says <a href="https://www.cleanegroup.org/staff/abbe-ramanan/" rel="noopener noreferrer" target="_blank">Abbe Ramanan</a>, project director at the Clean Energy Group, a Vermont-based nonprofit.</p><p>Ramanan explains that utilities often use “peaker plants” to cope with extra demand. They’re usually older, less efficient fossil-fuel plants which, because of their high cost to operate and carbon output, were due for retirement. But Ramanan says increased electricity demand <a href="https://www.eia.gov/todayinenergy/detail.php?id=61425" rel="noopener noreferrer" target="_blank">has kept them in service</a>.</p><p>Meta secured power for Hyperion by negotiating with Entergy, Louisiana’s electric utility, for construction of three new gas-turbine power plants. Two will be located near the Richland Parish site, while a third will be located in southeast Louisiana.</p><p>Entergy frames the new plants as a win for the state. “A core pillar of Entergy and Meta’s agreement is that Meta pays for the full cost of the utility infrastructure,” says <a href="https://www.linkedin.com/in/daniel-kline-068356ba/" rel="noopener noreferrer" target="_blank">Daniel Kline</a>, director of power-delivery planning and policy at Entergy. The utility expects that “customer bills will be lower than they otherwise would have been.” That would prove an exception, as <a href="https://www.bloomberg.com/graphics/2025-ai-data-centers-electricity-prices/?embedded-checkout=true" rel="noopener noreferrer" target="_blank">a recent report from Bloomberg found</a> electricity rates in regions with data centers are more likely to increase than in regions without.</p><h3>CO2</h3><br/><img alt="Diagram of CO2 molecule with black carbon and red oxygen atoms connected by lines." class="rm-shortcode" data-rm-shortcode-id="c9cf38ac7004d413b7fe5b8b577a3d3d" data-rm-shortcode-name="rebelmouse-image" id="3b1b0" loading="lazy" src="https://spectrum.ieee.org/media-library/diagram-of-co2-molecule-with-black-carbon-and-red-oxygen-atoms-connected-by-lines.png?id=65348689&width=980"/><p>Research <a href="https://www.nature.com/articles/s41893-025-01681-y" target="_blank">published in Nature</a> in 2025 projects that data-center emissions will range from 24 million to 44 million CO2-equivalent metric tonnes annually through 2030 in the United States alone. While some materials used in data centers, such as concrete, lead to significant emissions, the majority of these emissions will result from the high energy demands of AI servers.</p><p>Estimating the carbon emissions of Hyperion is difficult, as the project won’t be completed until 2030. Assuming that the three new natural gas plants that are planned for construction as part of the project produce emissions typical for their type, however, the plants could lead to full life-cycle emissions of between 4 million and 10 million metric tons of CO2 annually—roughly equivalent to the annual emissions of a country like <a href="https://www.worldometers.info/co2-emissions/co2-emissions-by-country/" target="_blank">Latvia</a>.</p><h3>Concrete</h3><br/><img alt="Silhouette of a cement truck on an orange background." class="rm-shortcode" data-rm-shortcode-id="060b1cd238b9de45274d6766069f3a14" data-rm-shortcode-name="rebelmouse-image" id="e6d68" loading="lazy" src="https://spectrum.ieee.org/media-library/silhouette-of-a-cement-truck-on-an-orange-background.png?id=65348696&width=980"/><p>Data centers are typically built from concrete, with steel used as a skeleton to reinforce and shape the concrete shell. While the foundation is often poured concrete, the walls and floors are most often built from prefabricated concrete panels that can span up to 23 meters. Floors use a reinforced T-shape, similar to a steel girder, measuring up to 1.2 meters across at its thickest point. The largest data centers include hundreds of these concrete panels.</p><p>The America Cement Association projects that the current surge in building<a href="https://mi.cement.org/PDF/Data_Center_Cement_Consumption.pdf" rel="noopener noreferrer" target="_blank"> will require 1 million tonnes of cement over the next three years</a>, though that’s still a tiny fraction of the overall cement industry,<a href="https://d9-wret.s3.us-west-2.amazonaws.com/assets/palladium/production/s3fs-public/media/files/mis-202507-cemen.pdf" rel="noopener noreferrer" target="_blank"> which weighed in at roughly 103 million tonnes in 2024</a>.</p><h3></h3><br/><p>The plants, which will generate a combined 2.26 GW, will use combined-cycle gas turbines that recapture waste heat from exhaust.<a href="https://www.ge.com/news/press-releases/ha-technology-now-available-industry-first-64-percent-efficiency" target="_blank"> This boosts thermal efficiency to 60 percent and beyond,</a> meaning more fuel is converted to useful energy. Simple-cycle turbines, by contrast, vent the exhaust, which lowers efficiency to around 40 percent.</p><p>Even so, total life-cycle emissions for the Hyperion plants could range from 4 million to over 10 million tonnes of CO2 each year, depending on how frequently the plants are put in use and the final efficiency benchmarks once built. On the high end, that’s as much CO2 as produced by over 2 million passenger cars. Fortunately, not all of Meta’s data centers take the same approach to power. The company has announced a plan to power Prometheus, a large data-center project in Ohio scheduled to come online before the end of 2026, <a href="https://about.fb.com/news/2026/01/meta-nuclear-energy-projects-power-american-ai-leadership/" target="_blank">with nuclear energy</a>.</p><p>But other big tech companies, spurred by the need to build data centers quickly, are taking a less efficient approach.</p><p>xAI’s Colossus 2, located in Memphis, is the most extreme example. <a href="https://www.climateandcapitalmedia.com/35-gas-turbines-no-permits-elon-musks-dirty-xai-secret/" rel="noopener noreferrer" target="_blank">The company trucked dozens of temporary gas-turbine generators to power the site</a> located in a suburban neighborhood. OpenAI, meanwhile, has gas turbines capable of generating up to 300 megawatts <a href="https://www.timesrecordnews.com/story/news/2025/10/14/water-electricity-concerns-addressed-by-stargate-data-center-leaders-in-abilene-texas/86585222007/" rel="noopener noreferrer" target="_blank">at its new Stargate data center in Abilene, Texas</a>, slated to open later in 2026. Both use simple-cycle turbines with a much lower efficiency rating than the combined-cycle plants Entergy will build to power Hyperion.</p><p>Demand for gas turbines is so intense, in fact, that <a href="https://www.spglobal.com/commodity-insights/en/news-research/latest-news/electric-power/052025-us-gas-fired-turbine-wait-times-as-much-as-seven-years-costs-up-sharply" rel="noopener noreferrer" target="_blank">wait times for new turbines are up to seven years</a>. Some data centers <a href="https://spectrum.ieee.org/ai-data-centers" target="_self">are turning toward refurbished jet engines</a> to obtain the turbines they need.</p><h2>AI Racks Tip the Scales</h2><p>The demand for new, reliable power is driven by the power-hungry GPUs inside modern AI data centers.</p><p>In January of 2025, Mark Zuckerberg announced in a post on Facebook that Meta planned to end 2025 <a href="https://techcrunch.com/2025/01/24/mark-zuckerberg-says-meta-will-have-1-3m-gpus-for-ai-by-year-end/" rel="noopener noreferrer" target="_blank">with at least 1.3 million GPUs in service</a>. OpenAI’s Stargate data center <a href="https://www.datacenterdynamics.com/en/news/openai-and-oracle-to-deploy-450000-gb200-gpus-at-stargate-abilene-data-center/" rel="noopener noreferrer" target="_blank">plans to use over 450,000 Nvidia GB200 GPUs</a>, and xAI’s Colossus 2, an expansion of Colossus, <a href="https://www.nextbigfuture.com/2025/09/xai-colossus-2-first-gigawatt-ai-training-data-center.html" rel="noopener noreferrer" target="_blank">is built to accommodate over 550,000 GPUs</a>.</p><p>GPUs, which remain by far the most popular for AI workloads, are bundled into human-scale monoliths of steel and silicon which, much like the data centers built to house them, are rapidly growing in weight, complexity, and power consumption.</p><h3>Memory</h3><br/><img alt="Outlined head with a microchip brain on blue background, symbolizing AI and technology." class="rm-shortcode" data-rm-shortcode-id="7cd8d3faff2d24fa591295b9efd9b1ba" data-rm-shortcode-name="rebelmouse-image" id="70372" loading="lazy" src="https://spectrum.ieee.org/media-library/outlined-head-with-a-microchip-brain-on-blue-background-symbolizing-ai-and-technology.png?id=65350865&width=980"/><p>In addition to raw compute performance, Nvidia GB200 NVL72 racks also require huge amounts of memory. An Nvidia GB200 NVL72 rack may include up to 13.4 terabytes of high-bandwidth memory, which implies a data-center campus at Hyperion’s scale will require at least several dozen petabytes.</p><p>The immense demand has sent memory prices soaring:<a href="https://wccftech.com/dram-prices-have-risen-by-a-whopping-172-this-year-alone/" rel="noopener noreferrer" target="_blank"> The price of DRAM, specifically DDR5, has increased 172 percent in 2025</a>.</p><h3>Power</h3><br/><img alt="" class="rm-shortcode" data-rm-shortcode-id="eaf0380400ba03875bf2ee910f35ab5d" data-rm-shortcode-name="rebelmouse-image" id="5bd7d" loading="lazy" src="https://spectrum.ieee.org/media-library/image.png?id=65350873&width=980"/><p>Hyperion is expected to use 5 gigawatts of power across 11 buildings, which works out to just under 500 megawatts per building, assuming each will be similar to its siblings. That’s enough to power roughly 4.2 million U.S. homes.</p><p>Just one Hyperion data center built at the Richland Parish site will consume twice as much power as xAI’s Colossus which, at the time of its completion in the summer of 2024, was among the largest data centers yet built.</p><h3></h3><br/><p>Nvidia’s <a href="https://www.nvidia.com/en-us/data-center/gb200-nvl72/" target="_blank">GB200 NVL72</a>—a rack-scale system—is currently a leading choice for AI data centers. A single GB200 rack contains 72 GPUs, 36 CPUs, and up to 17 terabytes of memory. It measures 2.2 meters tall, <a href="https://aivres.com/wp-content/uploads/KRS8000v3.1.pdf" target="_blank">tips the scales at up to </a>1,553 kilograms, and consumes about 120 kilowatts—as much as around 100 U.S. homes. And this, according to Nvidia, is just the beginning. The company anticipates future racks could <a href="https://www.tomshardware.com/tech-industry/nvidia-to-boost-ai-server-racks-to-megawatt-scale-increasing-power-delivery-by-five-times-or-more" target="_blank">consume up to a megawatt each</a>.</p><p><a href="https://www.linkedin.com/in/viktorpetik/?originalSubdomain=hr" target="_blank">Viktor Petik</a>, senior vice president of infrastructure solutions at<a href="https://www.vertiv.com/en-us/" rel="noopener noreferrer" target="_blank"> Vertiv</a>, says the rapid change in rack-scale AI systems has forced data centers to adapt. “AI racks consume far more power and weigh more than their predecessors,” says Petik. He adds that data centers must supply racks with multiple power feeds, without taking up extra space.</p><p>The new power demands from rack-scale systems have consequences that are reflected in the design of the data center—even its footprint.</p><p>In 2022 Meta broke ground on a new data center at a campus in Temple, Texas. According to <a href="https://semianalysis.com/" rel="noopener noreferrer" target="_blank">SemiAnalysis</a>, which studies AI data centers, construction began with the intent <a href="https://newsletter.semianalysis.com/p/datacenter-anatomy-part-1-electrical" rel="noopener noreferrer" target="_blank">to build the data center in an H-shaped configuration common to other Meta data centers</a>.</p><h3>LAND</h3><br/><img alt="Black location pin icon on orange background." class="rm-shortcode" data-rm-shortcode-id="a2b2e04f07bd0ed3f60e1f86029497af" data-rm-shortcode-name="rebelmouse-image" id="248cd" loading="lazy" src="https://spectrum.ieee.org/media-library/black-location-pin-icon-on-orange-background.png?id=65351137&width=980"/><h3></h3><br/><p>Meta CEO Mark Zuckerberg kicked off the buzz around Hyperion by saying it would cover a large chunk of Manhattan. Many took that to mean Hyperion would be a single building of that size, which isn’t correct. Hyperion will actually be a cluster of data centers—11 are currently planned—with over 370,000 square meters of floor space. That’s a lot smaller even than New York City’s Central Park, which covers 6 percent of Manhattan.</p><p>Meta has room to grow, however. The Richland Parish site spans 14.7 million square meters in total, which is about a quarter the area of Manhattan. And the 370,000 square meters of floor space Hyperion is expected to provide doesn’t include external infrastructure, such as the three new combined-cycle gas power plants Louisiana utility Entergy is building to power the project.</p><h3></h3><br/><img alt="Map with site layout and regional location in Louisiana, showing roads and distances." class="rm-shortcode" data-rm-shortcode-id="b0cc9253de57aefb96d39a9892c95fe5" data-rm-shortcode-name="rebelmouse-image" id="a41a4" loading="lazy" src="https://spectrum.ieee.org/media-library/map-with-site-layout-and-regional-location-in-louisiana-showing-roads-and-distances.png?id=65352088&width=980"/><h3></h3><br/><p><span>Construction was paused midway in December of 2022, however, </span><a href="https://www.datacenterdynamics.com/en/news/exclusive-after-meta-cancels-odense-data-center-expansion-other-projects-are-being-rescoped/" target="_blank">as part of a company-wide review of its data-center infrastructure</a><span>. Meta decided to knock down the structure it had built and start from scratch. The reasons for this decision were never made public, but analysts believe it was due to the old design’s inability to deliver sufficient electricity to new, power-hungry AI racks. Construction resumed in 2023.</span></p><p>Meta’s replacement ditches the H-shaped building for simple, long, rectangular structures, each flanked by rows of gas-turbine generators. While Meta’s plans are subject to change, Hyperion is currently expected to comprise 11 rectangular data centers, each packed with hundreds of thousands of GPUs, spread across the 13.6-square-kilometer Richland Parish campus.</p><h2>Cooling, and Connecting, at Scale</h2><p>Nvidia’s ultradense AI GPU racks are changing data centers not only with their weight, and power draw, but also with their intense cooling and bandwidth requirements.</p><p>Data centers traditionally use air cooling, but that approach has reached its limits. “Air as a cooling medium is inherently inferior,” says<a href="https://cde.nus.edu.sg/me/staff/lee-poh-seng/" target="_blank"> Poh Seng Lee</a>, head of <a href="https://blog.nus.edu.sg/coolestlab/" rel="noopener noreferrer" target="_blank">CoolestLAB</a>, a cooling research group at the National University of Singapore.</p><p>Instead, going forward, GPUs will rely on liquid cooling. However, that adds a new layer of complexity. “It’s all the way to the facilities level,” says Lee. “You need pumps, which we call a coolant distribution unit. The CDU will be connected to racks using an elaborate piping network. And it needs to be designed for redundancy.” On the rack, pipes connect to cold plates mounted atop every GPU; outside the data-center shell, pipes route through evaporation cooling units. Lee says retrofitting an air-cooled data center is possible but expensive.</p><p>The networking used by AI data centers is also changing to cope with new requirements. Traditional data centers were positioned near network hubs for easy access to the global internet. AI data centers, though, are more concerned with networks of GPUs.</p><p>These connections must sustain high bandwidth with impeccable reliability. Mark Bieberich, a vice president at network infrastructure company Ciena, says its latest fiber-optic transceiver technology,<a href="https://www.ciena.com/products/wavelogic/wavelogic-6" rel="noopener noreferrer" target="_blank"> WaveLogic 6</a>, can provide up to 1.6 terabytes per second of bandwidth per wavelength. A single fiber can support 48 wavelengths in total, and Ciena’s largest customers have hundreds of fiber pairs, placing total bandwidth in the thousands of terabits per second.</p><h3></h3><br/><img alt="a piece of land with a big platform in the middle." class="rm-shortcode" data-rm-shortcode-id="fb6adbcb1ff833934363d6f6ce9cf993" data-rm-shortcode-name="rebelmouse-image" id="63272" loading="lazy" src="https://spectrum.ieee.org/media-library/a-piece-of-land-with-a-big-platform-in-the-middle.jpg?id=65343457&width=980"/><p><span>This is a point where the scale of Meta’s Hyperion, and other large AI data centers, can be deceptive. It seems to imply the physical size of a single data center is what matters. But rather than being a single building,</span><a href="https://datacenters.atmeta.com/richland-parish-data-center/" target="_blank"> Hyperion is actually a set of buildings</a><span> connected by high-speed fiber-optics.</span></p><p>“Interconnecting data centers is absolutely essential,” says Bieberich. “You could think about it as one logical AI training facility, but with geographically distributed facilities.” Nvidia has taken to calling this “scale across,” to contrast it with the idea that data centers must “scale up” to larger singular buildings.</p><h2>The Big but Hazy Future</h2><p>The full scale of the challenges that face Hyperion, and other future AI data centers of similar scale, remain hazy. Nvidia has yet to introduce the rack-scale AI GPU systems it will host. How much power will it demand? What type of cooling will it require? How much bandwidth must be provided? These can only be estimated.</p><p>In the absence of details, the gravity of AI data-center design is pulled toward one certainty: It must be big. New data-center designers are rewriting their rule book to handle power, cooling, and network infrastructure at a scale that would’ve seemed ridiculous five years ago.</p><p>This innovation is fueled by big tech’s fat wallet, which shelled out tens of billions of dollars in 2025 alone, leading to<a href="https://hbr.org/2025/10/is-ai-a-boom-or-a-bubble" target="_blank"> questions about whether the spending is sustainable</a>. For the engineers in the trenches of data-center design, though, it’s viewed as an opportunity to make the impossible possible.</p><p> “I tell my engineers, this is peak. We’re being engineers. We’re being asked complicated questions,” says Stantec’s Carter. “We haven’t got to do that in a long time.” <span class="ieee-end-mark"></span></p><p><em>This article appears in the April 2026 print issue.</em></p>]]></description><pubDate>Tue, 24 Mar 2026 15:00:05 +0000</pubDate><guid>https://spectrum.ieee.org/5gw-data-center</guid><category>Ai</category><category>Power</category><category>Construction</category><category>Data-centers</category><category>Type-cover</category><dc:creator>Matthew S. Smith</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/construction-symbols-on-yellow-background.png?id=65356154&amp;width=980"></media:content></item><item><title>Transforming Data Science With NVIDIA RTX PRO 6000 Blackwell Workstation Edition</title><link>https://spectrum.ieee.org/nvidia-rtx-pro-6000-pny</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/computer-setup-with-a-monitor-displaying-forest-graphics-keyboard-mouse-and-a-sleek-cpu-design.png?id=65315285&width=1245&height=700&coordinates=0%2C91%2C0%2C92"/><br/><br/><p><em>This is a sponsored article brought to you by <a href="https://www.pny.com/" target="_blank">PNY Technologies</a>.</em></p>In today’s data-driven world, data scientists face mounting challenges in preparing, scaling, and processing massive datasets. Traditional CPU-based systems are no longer sufficient to meet the demands of modern AI and analytics workflows. <a href="https://www.pny.com/nvidia-rtx-pro-6000-blackwell-ws?iscommercial=true&utm_source=IEEE+Spectrum+Blog&utm_medium=RTX+PRO+6000+body&utm_campaign=Blackwell+Workstation&utm_id=RTX+PRO+6000" rel="noopener noreferrer" target="_blank">NVIDIA RTX PRO<sup>TM</sup> 6000 Blackwell Workstation Edition</a> offers a transformative solution, delivering accelerated computing performance and seamless integration into enterprise environments.<h2>Key Challenges for Data Science</h2><ul><li><strong>Data Preparation: </strong>Data preparation is a complex, time-consuming process that takes most of a data scientist’s time.</li><li><strong>Scaling: </strong>Volume of data is growing at a rapid pace. Data scientists may resort to downsampling datasets to make large datasets more manageable, leading to suboptimal results.</li><li><strong>Hardware: </strong>Demand for accelerated AI hardware for data centers and cloud service providers (CSPs) is exceeding supply. Current desktop computing resources may not be suitable for data science workflows.</li></ul><h2>Benefits of RTX PRO-Powered AI Workstations</h2><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition delivers ultimate acceleration for data science and AI workflows. These powerful and robust workstations enable real-time rendering, rapid prototyping, and seamless collaboration. With support for up to four <a href="https://www.pny.com/nvidia-rtx-pro-6000-blackwell-max-q?iscommercial=true&utm_source=IEEE+Spectrum+Blog&utm_medium=RTX+PRO+6000+Blackwell+Max-Q+body&utm_campaign=Blackwell+Workstation&utm_id=RTX+PRO+6000" rel="noopener noreferrer" target="_blank">NVIDIA RTX PRO 6000 Blackwell Max-Q Workstation Edition</a> GPUs, users can achieve data center-level performance right at their desk, making even the most demanding tasks manageable.</p><p class="shortcode-media shortcode-media-youtube"> <span class="rm-shortcode" data-rm-shortcode-id="61bf7564ac8304e10487689487367c94" style="display:block;position:relative;padding-top:56.25%;"><iframe frameborder="0" height="auto" lazy-loadable="true" scrolling="no" src="https://www.youtube.com/embed/jwxxgHsU1jA?rel=0" style="position:absolute;top:0;left:0;width:100%;height:100%;" width="100%"></iframe></span> <small class="image-media media-caption" placeholder="Add Photo Caption...">PNY is redefining professional computing with the ‪@NVIDIA‬ RTX PRO 6000 Blackwell Workstation Edition, the most powerful desktop GPU ever built. Engineered for unmatched compute power, massive memory capacity, and breakthrough performance, this cutting-edge solution delivers a quantum leap forward in workflow efficiency, enabling professionals to tackle the most demanding applications with ease.</small><small class="image-media media-photo-credit" placeholder="Add Photo Credit...">PNY</small></p><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition empowers data scientists to handle massive datasets, perform advanced visualizations, and support multi-user environments without compromise. It’s ideal for organizations scaling up their analytics or running complex models. NVIDIA RTX PRO 6000 Blackwell Workstation Edition is optimized for AI workflows, leveraging the NVIDIA AI software stack, including CUDA-X, and NVIDIA Enterprise software. These platforms enable zero-code-change acceleration for Python-based workflows and support over 100 AI-powered applications, streamlining everything from data preparation to model deployment.</p><p>Finally, NVIDIA RTX PRO 6000 Blackwell Workstation Edition offers significant advantages in security and cost control. By offloading compute from the data center and reducing reliance on cloud resources, organizations can lower expenses and keep sensitive data on-premises for enhanced protection.</p><h2>Accelerate Every Step of Your Workflow</h2><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition is designed to transform the entire data science pipeline, delivering end-to-end acceleration from data preparation to model deployment. With NVIDIA CUDA-X open-source data science cuDF library and other GPU-accelerated libraries, data scientists can process massive datasets at lightning speed, often achieving up to 50X faster performance compared to traditional CPU-based tools. This means tasks like cleaning data, managing missing values, and engineering features can be completed in seconds, not hours, allowing teams to focus on extracting insights and building better models.</p><p class="pull-quote">NVIDIA RTX PRO 6000 Blackwell Workstation Edition is designed to transform the entire data science pipeline, delivering end-to-end acceleration from data preparation to model deployment</p><p>Exploratory data analysis is elevated with advanced analytics and interactive visualizations, powered by NVIDIA CUDA-X and PyData libraries. These tools enable users to create expansive, responsive visualizations that enhance understanding and support critical decision-making. When it comes to model training, GPU-accelerated XGBoost slashes training times from weeks to minutes, enabling rapid iteration and faster time-to-market AI solutions.</p><p>NVIDIA RTX PRO 6000 Blackwell Workstation Edition streamlines collaboration and scalability. With NVIDIA AI Workbench, teams can set up projects, develop, and collaborate seamlessly across desktops, cloud platforms, and data centers. The unified software stack ensures compatibility and robustness, while enterprise-grade hardware maximizes uptime and reliability for demanding workflows.</p><p>By integrating these advanced capabilities, NVIDIA RTX PRO 6000 Blackwell Workstation Edition empowers data scientists to overcome bottlenecks, boost productivity, and drive innovation, making them an essential foundation for modern, enterprise-ready AI development.</p><h2>Performance Benchmarks</h2><p>NVIDIA’s cuDF library offers zero-code change acceleration for pandas, delivering up to 50X performance gains. For example, a join operation that takes nearly 5 minutes on CPU completes in just 14 seconds on GPU. Advanced group by operations drop from almost 4 minutes to just 4 seconds.</p><h2>Enterprise-Ready Solutions from PNY</h2><p class="shortcode-media shortcode-media-rebelmouse-image rm-float-left rm-resized-container rm-resized-container-25" data-rm-resized-container="25%" rel="float: left;" style="float: left;"> <img alt="Black PNY logo with stylized uppercase letters on a transparent background." class="rm-shortcode" data-rm-shortcode-id="247ffcd9e141f1fc61c5172c5440d97e" data-rm-shortcode-name="rebelmouse-image" id="170af" loading="lazy" src="https://spectrum.ieee.org/media-library/black-pny-logo-with-stylized-uppercase-letters-on-a-transparent-background.png?id=65315393&width=980"/></p><p>Available from leading OEM manufacturers, NVIDIA RTX PRO 6000 Blackwell Workstation Edition Series GPUs are specifically engineered to meet the rigorous demands of enterprise environments. These systems incorporate NVIDIA Connect-X networking, now available at PNY and a comprehensive suite of deployment and support tools, ensuring seamless integration with existing IT infrastructure.</p><p>Designed for scalability, the latest generation of workstations can tackle complex AI development workflows at scale for training, development, or inferencing. Enterprise-grade hardware maximizes uptime and reliability.</p><p><strong>To learn more about NVIDIA RTX PRO™ Blackwell solutions, </strong><strong>visit:</strong> <a href="https://www.pny.com/professional/software-solutions/blackwell-architecture?utm_source=IEEE+Spectrum+Blog&utm_medium=Blackwell+Desktop+GPUs+learn+more&utm_campaign=Blackwell+Workstation&utm_id=RTX+PRO+6000" target="_blank">NVIDIA RTX PRO Blackwell | PNY Pro | pny.com</a> or email <a href="mailto:gopny@pny.com" target="_blank">GOPNY@PNY.COM</a><strong></strong></p>]]></description><pubDate>Mon, 23 Mar 2026 13:00:04 +0000</pubDate><guid>https://spectrum.ieee.org/nvidia-rtx-pro-6000-pny</guid><category>Artificial-intelligence</category><category>Computing</category><category>Data-science</category><category>Gpu-acceleration</category><category>Ai-workstations</category><category>Nvidia</category><dc:creator>PNY Technologies</dc:creator><media:content medium="image" type="image/png" url="https://spectrum.ieee.org/media-library/computer-setup-with-a-monitor-displaying-forest-graphics-keyboard-mouse-and-a-sleek-cpu-design.png?id=65315285&amp;width=980"></media:content></item><item><title>Startups Bring Optical Metamaterials to AI Data Centers</title><link>https://spectrum.ieee.org/optical-metamaterials-ai-data-centers</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-hand-holding-a-microchip-between-thumb-and-forefinger.jpg?id=65322426&width=1245&height=700&coordinates=0%2C187%2C0%2C188"/><br/><br/><p><span>Light-warping physics made “invisibility cloaks” a possibility. Now two startups hope to harness the science underlying this advance to boost the bandwidth of data centers and speed artificial intelligence.</span></p><p>Roughly 20 years ago, scientists developed the <a href="https://www.science.org/doi/10.1126/science.1125907" target="_blank">first</a> <a href="https://www.science.org/doi/10.1126/science.1133628" target="_blank"> structures</a> capable of curving light around objects to conceal them. These are composed of optical <a href="https://spectrum.ieee.org/two-photon-lithography-3d-printing" target="_self">metamaterials</a>—materials with structures smaller than the wavelengths they are designed to manipulate, letting them bend light in unexpected ways.</p><p>The problem with optical cloaks? “There’s no market for them,” says Patrick Bowen, cofounder and CEO of photonic computing startup <a href="https://www.neurophos.com/" target="_blank">Neurophos</a> in Austin, Texas. For instance, each optical cloak typically works only on a single color of light instead of on all visible colors as you might want for stealth applications.</p><p>Now companies are devising more practical uses for the science behind cloaks, such as improving the switches that connect computers in data centers for AI and other cloud services. Increasingly, <a href="https://newsletter.semianalysis.com/p/google-apollo-the-3-billion-game" target="_blank">data centers are looking to use optical circuit switches </a>to overcome the bandwidth limits and power consumption of conventional electronic switches and networks that require converting data between light to electrons multiple times.</p><p class="ieee-inbody-related">RELATED:  <a href="https://spectrum.ieee.org/optical-interconnects-imec-silicon-photonics" target="_blank">Semiconductor Industry Closes in on 400 Gb/s Photonics Milestone</a></p><p>However, today’s optical switching technologies have drawbacks of their own. For instance, ones that depend on silicon photonics face problems with energy efficiency, while those that rely on <a href="https://spectrum.ieee.org/self-assembly" target="_self">microelectromechanical systems (MEMS)</a> can prove unreliable, says Sam Heidari, CEO of optical metasurface startup <a href="https://lumotive.com/" rel="noopener noreferrer" target="_blank">Lumotive</a> in Redmond, Wash.</p><p>Instead, <a href="https://www.nature.com/articles/s44287-024-00136-4" rel="noopener noreferrer" target="_blank">Lumotive has developed metamaterials with adjustable properties</a>. Its new microchip, which debuted 19 March, is covered with copper structures built using standard chipmaking techniques. Between these copper features are <a href="https://spectrum.ieee.org/metasurface-displays" target="_self">liquid crystal</a> elements. The structure of these elements are electronically programmable, just like in liquid crystal displays (LCDs), to alter the optical properties of the metamaterial chip.</p><p>The microchip can precisely steer, lens, shape, and split beams of light reflected off its surface. It can perform all the same functions as multiple optical components with no moving parts in a programmable way in real time, according to Lumotive. “Having no moving parts significantly improves reliability,” Heidari says.</p><p>“We had to go through a lot of R&D at the foundries to not only make our devices functional, but also commercially viable in terms of the right cost and right reliability,” Heidari says.</p><p>The company says its new chips are capable of manipulating not only the industry’s standard of 256 by 256 ports, but could scale up to 10,000 by 10,000. “We think this is game-changing for data centers,” Heidari says. Lumotive plans to launch its first optical switches at the end of 2026.</p><h2>Optical Computing With Metamaterials</h2><p>Similarly, Neurophos hopes its technology may be transformative for artificial intelligence. Since AI is proving energy hungry when run on conventional electronics, scientists are exploring <a href="https://spectrum.ieee.org/optical-neural-networks" target="_self">optical computing</a> as a low-power alternative by processing data with light instead of electrons.</p><p>However, optical processors in the works today are typically far too bulky to achieve a compute density competitive with the best modern electronic processors, Bowen says. Neurophos says it can use metamaterials to build optical modulators—the optical equivalent of a transistor—that are 1/10,000th the size of today’s designs using standard chipmaking processes. “It’s entirely CMOS,” Bowen says. “There are no exotic materials in it.”</p><p>When a laser beam encoding data shines on a Neurophos chip, the way in which each metamaterial element is configured alters the reflected beam to encode results from complex AI tasks. “We basically fit a 1,000- by-1,000 array of optical modulators on a tiny 5-by-5-millimeter area on a chip,” Bowen says. “If you wanted to do that with off-the-shelf silicon photonics, your chip would be a square meter in size.”</p><p>All in all, Bowen claims the Neurophos microchip will offer 50 times greater compute density and 50 times greater energy efficiency than Nvidia’s Blackwell-generation GPU. The company says that hyperscalers—the world’s biggest cloud service providers—will evaluate two upcoming proof-of-concept chips this year. Neurophos is targeting its first systems for early 2028, with production ramping mid-2028.</p>]]></description><pubDate>Thu, 19 Mar 2026 19:19:43 +0000</pubDate><guid>https://spectrum.ieee.org/optical-metamaterials-ai-data-centers</guid><category>Artificial-intelligence</category><category>Data-center</category><category>Optical-switch</category><category>Optical-computing</category><category>Metamaterial</category><category>Metamaterials</category><dc:creator>Charles Q. Choi</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-hand-holding-a-microchip-between-thumb-and-forefinger.jpg?id=65322426&amp;width=980"></media:content></item><item><title>ENIAC, the First General-Purpose Digital Computer, Turns 80</title><link>https://spectrum.ieee.org/eniac-80-ieee-milestone</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/wide-view-of-men-and-women-working-on-the-eniac-in-the-1940s-all-four-walls-from-floor-to-ceiling-host-different-pieces-of-t.jpg?id=65315846&width=1245&height=700&coordinates=0%2C187%2C0%2C188"/><br/><br/><p>Happy 80th anniversary, ENIAC! The <a href="https://penntoday.upenn.edu/news/penns-eniac-worlds-first-electronic-computer-turns-80" rel="noopener noreferrer" target="_blank">Electronic Numerical Integrator and Computer</a>, the first large-scale, general-purpose, programmable electronic digital computer, helped shape our world.</p><p>On 15 February 1946, ENIAC—developed in the <a href="https://facilities.upenn.edu/maps/locations/moore-school-building" rel="noopener noreferrer" target="_blank">Moore School of Electrical Engineering</a> at the <a href="https://www.upenn.edu/" rel="noopener noreferrer" target="_blank">University of Pennsylvania</a>, in Philadelphia—was publicly demonstrated for the first time. Although primitive by today’s standards, ENIAC’s purely electronic design and programmability were breakthroughs in computing at the time. ENIAC made high-speed, general-purpose computing practicable and laid the foundation for today’s machines.</p><p>On the eve of its unveiling, the <a href="https://www.war.gov/" rel="noopener noreferrer" target="_blank">U.S. Department of War</a> issued a<a href="https://americanhistory.si.edu/comphist/pr1.pdf" rel="noopener noreferrer" target="_blank">news release</a> hailing it as a new machine “expected to revolutionize the mathematics of engineering and change many of our industrial design methods.” Without a doubt, electronic computers have transformed engineering and mathematics, as well as practically every other domain, including politics and spirituality.</p><p>ENIAC’s success ushered the modern computing industry and laid the foundation for today’s digital economy. During the past eight decades, computing has grown from a niche scientific endeavor into an engine of economic growth, the backbone of billion-dollar enterprises, and a catalyst for global innovation. Computing has led to a chain of innovations and developments such as stored programs, semiconductor electronics, integrated circuits, networking, software, the Internet, and distributed large-scale systems.</p><h2>Inside the ENIAC</h2><p>The motivation for developing ENIAC was the <a href="https://www.pbs.org/wgbh/aso/databank/entries/dt45en.html" rel="noopener noreferrer" target="_blank">need for faster computation</a> during World War II. The U.S. military wanted to produce extensive artillery firing tables for field gunners to quickly determine settings for a specific weapon, a target, and conditions. Calculating the tables by hand took “<a href="https://cacm.acm.org/blogcacm/computers-were-originally-humans/" rel="noopener noreferrer" target="_blank">human computers</a>” several days, and the available mechanical machines were far too slow to meet the demand.</p><h3>80 Years of Electronic Computer Milestones </h3><br/><h4>1946</h4><p><a href="https://www.britannica.com/technology/ENIAC" rel="noopener noreferrer" target="_blank"><strong>ENIAC operational</strong></a></p><p>Birth of electronic computing</p><h4>1951</h4><p><a href="https://www.britannica.com/technology/UNIVAC" target="_blank"><strong>UNIVAC I</strong></a></p><p><a href="https://www.britannica.com/technology/UNIVAC" target="_blank"></a>Start of commercial computing</p><h4>1958</h4><p><a href="https://www.synopsys.com/glossary/what-is-integrated-circuit.html" target="_blank"><strong>Integrated circuit</strong></a></p><p>Foundation for modern computer hardware</p><h4>1964</h4><p><a href="https://www.ibm.com/history/system-360" rel="noopener noreferrer" target="_blank"><strong>IBM System/360</strong></a></p><p>Popular mainframe computer</p><h4>1970</h4><p><a href="https://en.wikipedia.org/wiki/PDP-11" rel="noopener noreferrer" target="_blank"><strong>Programmed Data Processor (PDP-11)</strong></a></p><p>Popular 16-bit minicomputer</p><h4>1971</h4><p><a href="https://computer.howstuffworks.com/microprocessor.htm" rel="noopener noreferrer" target="_blank"><strong>Intel 4004</strong></a></p><p>Beginning of the microprocessor and microcomputer era</p><h4>1975</h4><p><a href="https://en.wikipedia.org/wiki/Cray-1" rel="noopener noreferrer" target="_blank"><strong>Cray-1</strong></a></p><p>First supercomputer</p><h4>1977</h4><p><a href="https://www.stromasys.com/resources/vax-computer-systems-an-in-depth-guide/" rel="noopener noreferrer" target="_blank"><strong>VAX</strong></a></p><p>Popular 32-bit minicomputer</p><h4>1981</h4><p><a href="https://en.wikipedia.org/wiki/IBM_Personal_Computer" rel="noopener noreferrer" target="_blank"><strong>IBM PC</strong></a></p><p>Personal and small-business computing</p><h4>1989</h4><p><a href="https://home.cern/science/computing/birth-web" rel="noopener noreferrer" target="_blank"><strong>World Wide Web</strong></a></p><p>Digital communication, interaction, and transaction (e-commerce)</p><h4>2002</h4><p><a href="https://en.wikipedia.org/wiki/Amazon_Web_Services" rel="noopener noreferrer" target="_blank"><strong>Amazon Web Services</strong></a></p><p>Beginning of the cloud computing revolution</p><h4>2010</h4><p><a href="https://en.wikipedia.org/wiki/IPad" rel="noopener noreferrer" target="_blank"><strong>Apple iPad</strong></a></p><p>Handheld computer/tablet</p><h4>2010</h4><p><a href="https://www.ibm.com/think/topics/industry-4-0" rel="noopener noreferrer" target="_blank"><strong>Industry 4.0</strong></a></p><p>Delivered real-time decision-making, smart manufacturing, and logistics</p><h4>2016</h4><p><a href="https://www.livescience.com/55642-reprogrammable-quantum-computer-created.html" rel="noopener noreferrer" target="_blank"><strong>First reprogrammable quantum computer demonstrated</strong></a></p><p>Ignited interest in quantum computing</p><h4>2023</h4><p><a href="https://en.wikipedia.org/wiki/Generative_artificial_intelligence" rel="noopener noreferrer" target="_blank"><strong>Generative AI boom</strong></a></p><p>Widespread use of GenAI by individuals, businesses, and academia</p><h4>2026</h4><p><a href="https://penntoday.upenn.edu/news/penns-eniac-worlds-first-electronic-computer-turns-80" rel="noopener noreferrer" target="_blank"><strong>ENIAC’s 80th anniversary</strong></a></p><p>80 years of computing evolution</p><h3></h3><br/><p>In 1942 <a href="https://www.britannica.com/biography/John-Mauchly" target="_blank">John Mauchly</a>, an associate professor of electrical engineering at Penn’s Moore School, suggested using vacuum tubes to speed up computer calculations. Following up on his theory, the U.S. Army <a href="https://en.wikipedia.org/wiki/Ballistic_Research_Laboratory" target="_blank">Ballistic Research Laboratory</a>, which was responsible for providing artillery settings to soldiers in the field, commissioned Mauchly and his colleagues<a href="https://www.britannica.com/biography/J-Presper-Eckert-Jr" target="_blank"></a><a href="https://ethw.org/J._Presper_Eckert" rel="noopener noreferrer" target="_blank">J. Presper Eckert</a> and <a href="https://ethw.org/Adele_Katz_Goldstine" target="_blank">Adele Katz Goldstine</a>, to work on a new high-speed computer. Eckert was a lab instructor at Moore, and Goldstine became one of ENIAC’s programmers. It took them a year to design ENIAC and 18 months to build it.</p><p>ENIAC contained about 18,000 vacuum tubes, which were cooled by 80 air blowers, and measured 8 feet by 3 feet (2.44 meters by 0.91 meters), and was almost 100 feet long (30.5 meters). It filled a large 30 feet by 50 feet (9.14m by 15.24m) room and weighed 30 tons (30,000 kilograms or 30 megagrams). It consumed as much electricity as a small town.</p><p>Programming the machine was <a href="https://www.pbs.org/wgbh/aso/databank/entries/dt45en.html" target="_blank">difficult</a>. ENIAC did not have stored programs, so to reprogram the machine, operators manually reconfigured cables with switches and plugboards, a process that took several days.</p><p>By the 1950s, large universities either had acquired or built their own machines to rival ENIAC. The schools included <a href="https://www.cam.ac.uk/" rel="noopener noreferrer" target="_blank">Cambridge</a> (EDSAC), <a href="https://www.mit.edu/" rel="noopener noreferrer" target="_blank">MIT</a> (Whirlwind), and <a href="https://www.princeton.edu/" rel="noopener noreferrer" target="_blank">Princeton</a> (IAS). Researchers used the computers to model physical phenomena, solve mathematical problems, and perform simulations.</p><p>After almost nine years of operation, ENIAC officially was decommissioned on 2 October 1955.</p><p><a href="https://mitpress.mit.edu/9780262535175/eniac-in-action/" rel="noopener noreferrer" target="_blank"><em>ENIAC in Action: Making and Remaking the Modern Computer</em></a>, a book by <a href="https://uwm.edu/history/about/directory/haigh-thomas/" rel="noopener noreferrer" target="_blank">Thomas Haigh</a>, <a href="https://mitpress.mit.edu/author/mark-priestley-15374/" rel="noopener noreferrer" target="_blank">Mark Priestley</a>, and <a href="https://www.researchgate.net/scientific-contributions/Crispin-Rope-2045495041" rel="noopener noreferrer" target="_blank">Crispin Rope</a>,<em> </em>describes the design, construction, and testing processes and dives into its afterlife use. The book also outlines the complex relationship between ENIAC and its designers, as well as the revolutionary approaches to computer architecture.</p><p>In the early 1970s, there was a controversy over who invented the electronic computer and who would be assigned the patent. In 1973 <a href="https://en.wikipedia.org/wiki/Earl_R._Larson" rel="noopener noreferrer" target="_blank">Judge Earl Richard Larson</a> of U.S. District Court in Minnesota ruled in the <a href="https://en.wikipedia.org/wiki/Honeywell,_Inc._v._Sperry_Rand_Corp." rel="noopener noreferrer" target="_blank">Honeywell <em><em>v.</em></em> Sperry Rand</a> case that Eckert and Mauchly did not invent the automatic electronic digital computer but instead had derived their subject matter from a <a href="https://jva.cs.iastate.edu/operation.php" rel="noopener noreferrer" target="_blank">computer</a> prototyped in 1939 by <a href="https://history-computer.com/people/john-vincent-atanasoff-complete-biography/" rel="noopener noreferrer" target="_blank">John Vincent Atanasoff</a> and Clifford Berry at Iowa State College (now <a href="https://www.iastate.edu/" rel="noopener noreferrer" target="_blank">Iowa State University</a>). The ruling granted Atanasoff legal recognition as the inventor of the first electronic digital computer.</p><h2>IEEE’s ENIAC Milestone</h2><p>In 1987 IEEE<a href="https://ethw.org/Milestones:Electronic_Numerical_Integrator_and_Computer,_1946" rel="noopener noreferrer" target="_blank">designated ENIAC</a> as an IEEE Milestone, citing it as “a major advance in the history of computing” and saying the machine “established the practicality of large-scale electronic digital computers and strongly influenced the development of the modern, stored-program, general-purpose computer.”</p><p>The commemorative Milestone plaque is displayed at the Moore School, by the entrance to the classroom where ENIAC was built.</p><h3></h3><br/><p>“The ENIAC legacy heralded the computer age, transforming not only science and industry but also education, research, and human communication and interaction.”</p><h3></h3><br/><p><br/></p><p>A <a href="https://ieeexplore.ieee.org/document/476557" rel="noopener noreferrer" target="_blank">paper on the machine</a>, published in 1996 in <a href="https://ieeexplore.ieee.org/document/476557" rel="noopener noreferrer" target="_blank"><em>IEEE Annals of the History of Computing</em></a> and available in the <a href="https://ieeexplore.ieee.org/document/6461145" rel="noopener noreferrer" target="_blank">IEEE Xplore Digital Library</a>, is a valuable source of technical information.</p><p>“<a href="https://www.computer.org/csdl/magazine/an/2006/02/man2006020004/13rRUB6Sq2p" rel="noopener noreferrer" target="_blank">The Second Life of ENIAC</a><em>,”</em> an article published in the annals in 2006, covers a lesser-known chapter in the machine’s history, about how it evolved from a static system—configured and reconfigured through laborious cable plugging—into a precursor of today’s stored-program computers.</p><p>A classic <a href="https://www2.seas.gwu.edu/~mfeldman/csci1030/summer08/eniac2.pdf" rel="noopener noreferrer" target="_blank">history paper on ENIAC</a> was published in the December 1995 <a href="https://technologyandsociety.org/" rel="noopener noreferrer" target="_blank"><em>IEEE Technology and Society Magazine</em></a>.</p><p>The IEEE <a href="https://spectrum.ieee.org/ebooks/ieee-anniversary-book/" target="_self"><em>Inspiring Technology: 34 Breakthroughs</em></a> book, published in 2023, features an ENIAC chapter.</p><h2>The women behind ENIAC</h2><p>One of the most remarkable aspects of the ENIAC story is the pivotal role women played, according to the book <a href="https://www.amazon.com/Proving-Ground-Untold-Programmed-Computer/dp/1538718286" rel="noopener noreferrer" target="_blank"><em>Proving Ground: The Untold Story of the Six Women Who Programmed the World’s First Modern Computer</em></a><em>, </em>highlighted in an <a href="https://spectrum.ieee.org/the-women-behind-eniac" target="_self">article</a> in <a href="https://spectrum.ieee.org/the-institute/" target="_self"><em>The Institute</em></a>. There were no “programmers” at that time; only schematics existed for the computer. Six women, known as the ENIAC 6, became the machine’s first programmers.</p><p>The ENIAC 6 were <a href="https://en.wikipedia.org/wiki/Kathleen_Antonelli" rel="noopener noreferrer" target="_blank">Kathleen Antonelli</a>, <a href="https://en.wikipedia.org/wiki/Jean_Bartik" rel="noopener noreferrer" target="_blank">Jean Bartik</a>, <a href="https://ethw.org/Betty_Holberton" rel="noopener noreferrer" target="_blank">Betty Holberton</a>, <a href="https://ethw.org/Marlyn_Meltzer" rel="noopener noreferrer" target="_blank">Marlyn Meltzer</a>, <a href="https://ethw.org/Frances_Spence" rel="noopener noreferrer" target="_blank">Frances Spence</a>, and <a href="https://ethw.org/Ruth_Teitelbaum" rel="noopener noreferrer" target="_blank">Ruth Teitelbaum</a>.</p><p>“These six women found out what it took to run this computer, and they really did incredible things,” a Penn professor, <a href="https://www.cis.upenn.edu/~mitch/" rel="noopener noreferrer" target="_blank">Mitch Marcus</a>, said in a <a href="https://www.phillyvoice.com/70-years-ago-six-philly-women-eniac-digital-computer-programmers/" rel="noopener noreferrer" target="_blank">2006 PhillyVoice article</a>. Marcus teaches in Penn’s computer and information science department.</p><p>In 1997 all six female programmers were<a href="https://www.witi.com/halloffame/298369/ENIAC-Programmers-Kathleen---/" rel="noopener noreferrer" target="_blank">inducted</a> into the <a href="https://www.witi.com/halloffame/" rel="noopener noreferrer" target="_blank">Women in Technology International Hall of Fame</a>, in Los Angeles.</p><p>Two other women contributed to the programming. Goldstine wrote ENIAC’s five-volume manual, and <a href="https://en.wikipedia.org/wiki/Kl%C3%A1ra_D%C3%A1n_von_Neumann" rel="noopener noreferrer" target="_blank">Klára Dán von Neumann</a>, wife of <a href="https://ethw.org/John_von_Neumann" rel="noopener noreferrer" target="_blank">John von Neumann</a>, helped train the programmers and debug and verify their code.</p><p>To honor the<a href="https://www.computer.org/volunteering/awards/pioneer/about-women-of-eniac" rel="noopener noreferrer" target="_blank">women of ENIAC</a>, the <a href="https://www.computer.org/" rel="noopener noreferrer" target="_blank">IEEE Computer Society</a> established the annual<a href="https://www.computer.org/volunteering/awards/pioneer" rel="noopener noreferrer" target="_blank">Computer Pioneer Award</a> in 1981. Eckert and Mauchly were among the award’s first recipients. In 2008 Bartik was honored with the award. Nominations are open to all professionals, regardless of gender.</p><h2>An ENIAC replica</h2><p>Last year a group of 80 autistic students, ages 12 to 16, from<a href="https://www.psacademyarizona.com/" rel="noopener noreferrer" target="_blank">PS Academy Arizona</a>, in Gilbert, <a href="https://www.msn.com/en-us/news/technology/how-80-autistic-students-built-an-amazing-replica-of-the-ginormous-eniac-computer/ar-AA1UMKKE" rel="noopener noreferrer" target="_blank">recreated the ENIAC</a> using 22,000 custom parts. It took the students almost six months to assemble.</p><p>A ceremony was held in January to display their creation. The full-scale <a href="https://www.theregister.com/2026/01/21/eniac_model_build/" rel="noopener noreferrer" target="_blank">replica features</a> actual-size panels made from layered cardboard and wood. Although all electronic components are simulated, they are not electrically active. The machine, illuminated by hundreds of LEDs, is accompanied by a soundtrack that simulates the deep hum of ENIAC’s transformers and the rhythmic clicking of relays.</p><p><strong></strong></p><h3></h3><br><img alt="A white woman using a computer-adding machine in the 1940\u2019s. The device resembles a bulky typewriter and prints large stacks of paper with tabulated answers." class="rm-shortcode" data-rm-shortcode-id="fea0fb9da93e75542fd5b85964251c33" data-rm-shortcode-name="rebelmouse-image" id="36a08" loading="lazy" src="https://spectrum.ieee.org/media-library/a-white-woman-using-a-computer-adding-machine-in-the-1940-u2019s-the-device-resembles-a-bulky-typewriter-and-prints-large-stack.jpg?id=65315890&width=980"/><h3></h3><br/><p>“Every major unit, accumulators, function tables, initiator, and master programmer is present and placed exactly where it was on the original machine,” Tom Burick, the teacher who mentored the project, said at the ceremony.</p><p>The replica, still on display at the school, is expected to be moved to a more permanent spot in the near future.</p><h2>ENIAC’s legacy</h2><p>ENIAC’s significance is both technical and symbolic. Technically, it marks the beginning of the chain of innovations that created today’s computational infrastructure. Symbolically, it made governments, militaries, universities, and industry view computation as a tool for improvement and for innovative applications that had previously been impossible. It marked a tectonic shift in the way humans approach problem-solving, modeling, and scientific reasoning.</p><p>The ENIAC legacy heralded the computer age, transforming not only science and industry but also education, research, and human communication and interaction.</p><p>As Eckert is reported to have said, “There are two epochs in computer history: Before ENIAC and After ENIAC.”</p><h2>Coevolution of programming languages</h2><p>The remarkable evolution of computer hardware during the past 80 years has been sparked by advances in programming languages—the essential drivers of computing.</p><p>From the manual rewiring of ENIAC to the orchestration of intelligent, distributed systems, programming languages have steadily evolved to make computers more powerful, expressive, and accessible.</p><h3>Lessons From Computing’s Remarkable Journey</h3><br/><p>Computing history teaches us that flexibility, accessibility, collaboration, sound governance, and forward thinking are essential for sustained technological progress. In a <a href="https://cacm.acm.org/blogcacm/what-past-computing-breakthroughs-teach-us-about-ai/" target="_blank">recent <em><em>Communications of the ACM</em></em> article</a>, <a href="https://www.linkedin.com/in/richa28gupta/" target="_blank">Richa Gupta</a> identified four historic shifts that led to computing’s rapid, transformative progress:</p><ol><li>Programmable machines taught us that flexibility is key; technologies that adapt and are repurposed scale better.</li><li>The Internet showed that connection and standard protocols drive explosive growth but also bring new risks such as data security issues, invasion of privacy, and misuse.</li><li>Personal computers illustrated that accessibility and usability matter more than raw power. When nonexperts can use a tool easily, adoption rises.</li><li>The open-source movement revealed that collaborative innovation accelerates growth and helps spot problems early.</li></ol></br><h2>Predictions for computing in the decades ahead</h2><p>The evolution of computing will continue along multiple trajectories, with the emphasis moving from generalization to specialization (for AI, graphics, security, and networking), from monolithic system design to modular integration, and from performance-centric metrics alone to energy efficiency and sustainability as primary objectives.</p><p>Increasingly, security will be built into hardware by design. Computing paradigms will expand beyond traditional deterministic models to embrace probabilistic, approximate, and hybrid approaches for certain tasks.</p><p>Those developments will usher in a new era of computing and a new class of applications.</p><p><em>This article was updated on 22 April 2026.</em></p>]]></description><pubDate>Wed, 18 Mar 2026 18:00:05 +0000</pubDate><guid>https://spectrum.ieee.org/eniac-80-ieee-milestone</guid><category>Ieee-history</category><category>Eniac</category><category>Computing</category><category>Computers</category><category>History-of-technology</category><category>Type-ti</category><dc:creator>San Murugesan</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/wide-view-of-men-and-women-working-on-the-eniac-in-the-1940s-all-four-walls-from-floor-to-ceiling-host-different-pieces-of-t.jpg?id=65315846&amp;width=980"></media:content></item><item><title>Wanted: Europe’s Missing Cloud Provider</title><link>https://spectrum.ieee.org/europe-cloud-sovereignty</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/abstract-pixelation-of-the-european-union-s-flag.jpg?id=65298877&width=1245&height=700&coordinates=0%2C156%2C0%2C157"/><br/><br/><p>Looming over the <a href="https://spectrum.ieee.org/free-space-optical-link-taara" target="_self">internet lasers</a> and <a href="https://www.pcmag.com/news/hands-on-with-oukitel-wp63-mwc-2026" rel="noopener noreferrer" target="_blank">firestarting phones</a> companies were touting at Mobile World Congress in Barcelona this month, was a more nebulous but much larger announcement: a pan-European cloud called <a href="https://www.euronews.com/next/2026/03/03/europe-unites-to-build-sovereign-cloud-and-ai-infrastructure-to-stop-reliance-on-us" rel="noopener noreferrer" target="_blank">EURO-3C</a>.</p><p>EURO-3C’s backers—Spanish telecoms giant Telefónica, dozens of other European companies, and the European Commission (EC)—aim to fill a gap. U.S.-based cloud giants dominate in the EU, and European policymakers want their growing portfolio of digital government services on a “sovereign cloud” under full EU control.</p><p>But the EU lacks a real equivalent to the likes of AWS or Microsoft Azure. Indeed, any effort to build one will inevitably run up against the same U.S. cloud giants.</p><p>Just four U.S.-based hyperscalers—AWS, Microsoft Azure, Google Cloud, and IBM Cloud—together account for<a href="https://www.ceps.eu/disk-backup-to-the-cloud-is-a-gaping-vulnerability-in-the-eus-security/" rel="noopener noreferrer" target="_blank"> some 70 percent of EU cloud services</a>. This is despite the fact that the 2018 U.S. <a href="https://en.wikipedia.org/wiki/CLOUD_Act" rel="noopener noreferrer" target="_blank">CLOUD Act</a> allows U.S. federal law enforcement—at least in theory—to compel U.S.-based firms to hand over data that’s stored abroad. </p><h2>Who Do You Trust?</h2><p>But those hypothetical risks to digital services have become more real as transatlantic relations have soured under the second Trump administration. The U.S. has <a href="https://www.cbc.ca/news/politics/greenland-us-trump-canada-governor-general-mary-simon-9.7119074" rel="noopener noreferrer" target="_blank">openly threatened</a> to invade an EU member state and <a href="https://euobserver.com/19745/eu-rejects-us-claims-of-censorship-over-tech-rules-after-visa-bans/" rel="noopener noreferrer" target="_blank">sanctioned</a> a European Commissioner for passing legislation the White House dislikes. </p><p>After the White House sanctioned the Netherlands-based International Criminal Court in February 2025, Court staffers <a href="https://apnews.com/article/icc-trump-sanctions-karim-khan-court-a4b4c02751ab84c09718b1b95cbd5db3" rel="noopener noreferrer" target="_blank">claimed</a> Microsoft locked the Court’s chief prosecutor out of his email (Microsoft<a href="https://www.politico.eu/article/microsoft-did-not-cut-services-international-criminal-court-president-american-sanctions-trump-tech-icc-amazon-google/" rel="noopener noreferrer" target="_blank"> has denied this</a>). Around the same time, the U.S. <a href="https://kyivindependent.com/us-threatens-to-shut-off-starlink-if-ukraine-wont-sign-minerals-deal-sources-tell-reuters/" rel="noopener noreferrer" target="_blank">reportedly threatened</a> to sever EU ally Ukraine’s access to crucial Starlink satellite internet as leverage during trade negotiations.</p><p>“The geopolitical risk isn’t just the most extreme form of a doomsday ‘kill switch’ where Washington turns off Europe’s internet,” says <a href="https://fermigier.com/" rel="noopener noreferrer" target="_blank">Stéfane Fermigier</a> of <a href="https://euro-stack.com/pages/about" rel="noopener noreferrer" target="_blank">EuroStack</a>, an industry group that supports European digital independence. “It is the selective degradation of services and a total lack of retaliatory leverage.”</p><p>What, then, is the EU to do? <a href="https://blog.datacenter-paris.com/2026/01/24/liste-des-datacenters-secnumcloud-en-france-hebergement-souverain-pour-donnees-sensibles/" rel="noopener noreferrer" target="_blank">France</a> offers an example. Even before 2025, France implemented <a href="https://www.spscommerce.com/eur/blog/what-is-secnumcloud-and-does-my-company-need-to-qualify/" rel="noopener noreferrer" target="_blank">harsh restrictions</a> on non-EU cloud providers in public services—providers must locate data in the EU, rely on EU-based staff, and may not have majority non-EU shareholders. Now, EU policymakers are following France’s lead.</p><p>In October 2025, the EC issued a two-part <a href="https://commission.europa.eu/document/09579818-64a6-4dd5-9577-446ab6219113_en" rel="noopener noreferrer" target="_blank">framework</a> for judging cloud providers bidding for public-sector contracts. In the first part, the framework lays out a sort of sovereignty ladder. The more that a provider is subject to EU law, the higher its sovereignty level on this ladder. Any prospective bidder must first meet a certain level, depending on the tender.</p><p>Qualifying bidders then move to the second part, where their “sovereignty” is scored in more detail. Using too much proprietary software; over-relying on supply chains from outside the EU; having non-EU support staff; liability to non-EU laws like the CLOUD Act: All hurt a bidder’s score. </p><p>The framework was created for <a href="https://commission.europa.eu/news-and-media/news/commission-moves-forward-cloud-sovereignty-eur-180-million-tender-2025-10-10_en" rel="noopener noreferrer" target="_blank">one tender</a>, but observers say it sets a major precedent. Cloud providers bidding for state contracts across Europe may need to follow it, and it may influence legislation on both national and EU-wide levels.</p><h2>A Question of Scale</h2><p>Who, then, will receive high marks? At the moment, the answer is not simple. The EU cloud scene is quite fragmented. Numerous modest EU providers offer “sovereign cloud” services—such as Deutsche Telekom’s T-Systems, OVHcloud, and Scaleway—but <a href="https://onlinelibrary.wiley.com/doi/full/10.1002/poi3.358" rel="noopener noreferrer" target="_blank">none are on the scale</a> of AWS or Google Cloud.</p><p>Inertia is on the side of the U.S. cloud giants, which can invest in their infrastructure and services on a far grander scale than their European counterparts. Some U.S. providers <a href="https://aws.amazon.com/blogs/security/aws-european-sovereign-cloud-achieves-first-compliance-milestone-soc-2-and-c5-reports-plus-seven-iso-certifications/" rel="noopener noreferrer" target="_blank">now offer</a> cloud services they say comply with the Commission’s “cloud sovereignty” demands.</p><p>Some European observers, like EuroStack, <a href="https://euro-stack.com/blog/2025/10/cloud-sovereignty-framework-comparison" rel="noopener noreferrer" target="_blank">say</a> such promises are hollow so long as a provider’s parent company is subject to the likes of the CLOUD Act and loopholes in the Commission’s process remain open. An AWS spokesperson told <em>IEEE </em><em>Spectrum</em> it had not disclosed any non-US enterprise or government data to the U.S. government under the CLOUD Act; a Google spokesperson said that its most sensitive EU offerings “are subject to local laws, not U.S. law”.</p><p>Even if a project like EURO-3C can offer a large-scale alternative, the U.S. cloud giants have another sort of inertia. Many developers—and many public purchasers of their services—will need convincing to leave behind a familiar environment.</p><p>“If you look at AWS, you look at Google, they’ve created some super technology. It’s very convenient, it’s easy to use,” says <a href="https://nl.linkedin.com/in/arnoldjuffer" rel="noopener noreferrer" target="_blank">Arnold Juffer</a>, CEO of the Netherlands-based cloud provider <a href="https://nebul.com/" rel="noopener noreferrer" target="_blank">Nebul</a>. “Once you’re in that platform, in that ecosystem, it’s very hard to get out.”</p><p><a href="https://bisi.org.uk/martyna-chmura" rel="noopener noreferrer" target="_blank">Martyna Chmura</a>, an analyst at the Bloomsbury Intelligence and Security Institute, a London-based think tank, sees some EU developers taking a mixed approach. “Many organizations are already moving toward multicloud setups, using European or sovereign providers for sensitive workloads while still relying on hyperscalers for certain services,” she says.</p><p>In that case, the EU’s top-down demands may encourage developers to use EU providers for sensitive applications—like government services, transport, autonomous vehicles, and some industrial automation—even if it’s inconvenient in the short term, or if it causes even more fragmentation of the EU cloud scene. “Running systems across different platforms can increase integration costs and make security and data governance more complicated. In some cases, organisations could lose some of the efficiency and cost advantages that come from using large hyperscale platforms,” Chmura says.</p><p>“Overall, the EU appears willing to accept some of these trade-offs,” Chmura says.</p>]]></description><pubDate>Tue, 17 Mar 2026 11:00:06 +0000</pubDate><guid>https://spectrum.ieee.org/europe-cloud-sovereignty</guid><category>Cloud-computing</category><category>Data-security</category><category>Data-privacy</category><dc:creator>Rahul Rao</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/abstract-pixelation-of-the-european-union-s-flag.jpg?id=65298877&amp;width=980"></media:content></item><item><title>With Nvidia Groq 3, the Era of AI Inference Is (Probably) Here</title><link>https://spectrum.ieee.org/nvidia-groq-3</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/a-man-in-all-black-presents-in-front-of-a-large-screen-which-compares-a-large-rectangular-chip-labelled-rubin-gpu-with-a-square.jpg?id=65298681&width=1245&height=700&coordinates=0%2C156%2C0%2C157"/><br/><br/><p>This week, over 30,000 people are descending upon San Jose, Calif., to attend<a href="https://www.nvidia.com/gtc/" rel="noopener noreferrer" target="_blank">Nvidia GTC</a>, the so-called Superbowl of AI—a nickname that may or may not have been coined by Nvidia. At the main event Jensen Huang, Nvidia CEO, took the stage to announce (among other things) a new line of<a href="https://spectrum.ieee.org/nvidia-rubin-networking" target="_self">next-generation Vera Rubin</a> chips that represent a first for the GPU giant: a chip designed specifically to handle AI inference. The Nvidia Groq 3 language processing unit (LPU) incorporates intellectual property Nvidia<a href="https://groq.com/newsroom/groq-and-nvidia-enter-non-exclusive-inference-technology-licensing-agreement-to-accelerate-ai-inference-at-global-scale" rel="noopener noreferrer" target="_blank">licensed</a> from the startup Groq last Christmas Eve for US $20 billion.</p><p>“Finally, AI is able to do productive work, and therefore the inflection point of inference has arrived,” Huang told the crowd. “AI now has to think. In order to think, it has to inference. AI now has to do; in order to do, it has to inference.”</p><p>Training and inference tasks have distinct computational requirements. While training can be done on huge amounts of data at the same time and can take weeks, inference must be run on a user’s query when it comes in. Unlike training, inference doesn’t require running costly<a href="https://spectrum.ieee.org/what-is-deep-learning/backpropagation" target="_self">backpropagation</a>. With inference, the most important thing is low latency—users expect the chatbot to answer quickly, and for thinking or reasoning models, inference runs many times before the user even sees an output.</p><p>Over the past few years, inference-specific chip startups were experiencing a sort of Cambrian explosion, with different companies exploring distinct approaches to speed up the task. The startups include<a href="https://www.d-matrix.ai/" rel="noopener noreferrer" target="_blank">D-matrix</a>, with digital in-memory compute;<a href="https://www.etched.com/" rel="noopener noreferrer" target="_blank">Etched</a>, with an ASIC for transformer inference;<a href="https://rain.ai/" rel="noopener noreferrer" target="_blank">RainAI</a>, with neuromorphic chips;<a href="https://en100.enchargeai.com/" rel="noopener noreferrer" target="_blank">EnCharge</a>, with analog in-memory compute;<a href="https://www.tensordyne.ai/" rel="noopener noreferrer" target="_blank">Tensordyne</a>, with logarithmic math to make AI computations more efficient;<a href="https://furiosa.ai/" rel="noopener noreferrer" target="_blank">FuriosaAI</a>, with hardware optimized for tensor operation rather than vector-matrix multiplication, and others.</p><p>Late last year, it looked like Nvidia had picked one of the winners among the crop of inference chips when it announced its deal with Groq. The Nvidia Groq 3 LPU reveal came a mere two and a half months after, highlighting the urgency of the growing inference market.</p><h2>Memory bandwidth and data flow</h2><p>Groq’s approach to accelerating inference relies on interleaving processing units with memory units on the chip. Instead of relying on high-bandwidth memory (HBM) situated next to GPUs, it leans on SRAM memory integrated within the processor itself. This design greatly simplifies the flow of data through the chip, allowing it to proceed in a streamlined, linear fashion.</p><p>“The data actually flows directly through the SRAM,”<a href="https://www.linkedin.com/in/markheaps/" rel="noopener noreferrer" target="_blank">Mark Heaps</a> said at the Supercomputing conference in 2024. Heaps was a chief technology evangelist at Groq at the time and is now director of developer marketing at Nvidia. “When you look at a multicore GPU, a lot of the instruction commands need to be sent off the chip, to get into memory and then come back in. We don’t have that. It all passes through in a linear order.”</p><p>Using SRAM allows that linear data flow to happen exceptionally fast, leading to the low latency required for inference applications. “The LPU is optimized strictly for that extreme low latency token generation,” says<a href="https://www.linkedin.com/in/ian-buck-19201315/" rel="noopener noreferrer" target="_blank">Ian Buck</a>, VP and general manager of hyperscale and high-performance computing at Nvidia.</p><p>Comparing the Rubin GPU and Groq 3 LPU side by side highlights the difference. The Rubin GPU has access to a whopping 288 gigabytes of HBM and is capable of 50 quadrillion floating-point operations per second (petaFLOPS) of 4-bit computation. The Groq 3 LPU contains a mere 500 megabytes of SRAM memory and is capable of 1.2 petaFLOPS of 8-bit computation. On the other hand, while the Rubin GPU has a memory bandwidth of 22 terabytes per second, at 150 TB/s the Groq 3 LPU is seven times as fast. The lean, speed-focused design is what allows the LPU to excel at inference.</p><p>The new inference chip underscores the ongoing trend of AI adoption, which shifts the computational load from just building ever bigger models to actually using those models at scale. “Nvidia’s announcement validates the importance of SRAM-based architectures for large-scale inference, and no one has pushed SRAM density further than d-Matrix,” says d-Matrix CEO Sid Sheth. He’s betting that data center customers will want a variety of processors for inference. “The winning systems will combine different types of silicon and fit easily into existing data centers alongside GPUs.”</p><p>Inference-only chips may not be the only solution. Late last week, <a href="https://press.aboutamazon.com/aws/2026/3/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud" rel="noopener noreferrer" target="_blank">Amazon Web Services</a> said that it will deploy a new kind of inferencing system in its data centers. The system is a combination of AWS’s Tranium <a href="https://spectrum.ieee.org/amazon-ai" target="_self">AI accelerator </a>and <a href="https://spectrum.ieee.org/cerebras-chip-cs3" target="_self">Cerebras Systems’ third generation computer CS-3</a>, which is built around the <a href="https://spectrum.ieee.org/cerebrass-giant-chip-will-smash-deep-learnings-speed-barrier" target="_self">largest single chip</a> ever made. The two-part system is meant to take advantage of a technique called inference disaggregation. It separates inference into two parts—processing the prompt, called prefill, and generating the output, called decode. Prefill is inherently parallel, computationally intensive, and doesn’t need much memory bandwidth, while decode is a more serial process that needs a lot of memory bandwidth. Cerebras has maximized the memory bandwidth issue by building 44 GB of SRAM on its chip connected by a 21 PB/s network. </p><p><span>Nvidia, too, intends to take advantage of inference disaggregation in its new compute rack, called the Nvidia <a href="https://developer.nvidia.com/blog/inside-nvidia-groq-3-lpx-the-low-latency-inference-accelerator-for-the-nvidia-vera-rubin-platform/" target="_blank">Groq 3 LPX</a>. Each tray within the rack will house 8 Groq 3 LPUs. The LPX will split the inference task with a <a href="https://www.nvidia.com/en-us/data-center/vera-rubin-nvl72/" target="_blank">Vera Rubin NVL72</a>, Nvidia’s existing GPU and CPU rack.</span> The prefill and the more computationally intensive parts of the decode are done on Vera Rubin, while the final part is done on the Groq 3 LPU, leveraging the strengths of each chip. “We’re in volume production now,” Huang said.</p><p><strong>Correction on 4/8/26: </strong>a previous version of this article incorrectly stated that the Nvidia Groq 3 LPX contains a Vera Rubin chip in each tray. In fact, each tray contains 8 Groq 3 LPUs and no Vera Rubins, but the whole rack is designed to work in concert with an NVL72 rack, which houses Vera Rubin chips. </p><p><em>This article appears in the May 2026 print issue as “<span>The Era of AI Inference Is Almost Here</span>.”</em></p>]]></description><pubDate>Mon, 16 Mar 2026 21:04:33 +0000</pubDate><guid>https://spectrum.ieee.org/nvidia-groq-3</guid><category>Inferencing</category><category>Nvidia</category><category>Gpus</category><category>Processors</category><category>Ai</category><dc:creator>Dina Genkina</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/a-man-in-all-black-presents-in-front-of-a-large-screen-which-compares-a-large-rectangular-chip-labelled-rubin-gpu-with-a-square.jpg?id=65298681&amp;width=980"></media:content></item><item><title>Intel Demos Chip to Compute With Encrypted Data</title><link>https://spectrum.ieee.org/fhe-intel</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/overhead-view-of-intel-s-computing-chip-called-heracles.jpg?id=65174073&width=1245&height=700&coordinates=0%2C156%2C0%2C157"/><br/><br/><div class="ieee-summary"><h2>Summary</h2><ul><li><a href="#fhe">Fully homomorphic encryption (FHE)</a> allows computing on encrypted data without decryption, but it’s currently slow on standard CPUs and GPUs.</li><li>Intel’s Heracles chip accelerates FHE tasks up to <a href="#faster">5,000 times as fast as</a> top Intel server CPUs.</li><li>Heracles uses a <a href="#heracles">3-nanometer FinFET technology and high-bandwidth memory</a>, enabling efficient encrypted computing at scale.</li><li>Startups and Intel are <a href="#commercial">racing to commercialize FHE accelerators</a>, with potential applications in AI and secure data processing.</li></ul></div><p><span>Worried that your latest ask to a cloud-based AI reveals a bit too much about you? Want to know your genetic risk of disease without revealing it to the services that compute the answer?</span></p><p>There is a way to do computing on encrypted data without ever having it decrypted. It’s called <a href="https://spectrum.ieee.org/homomorphic-encryption" target="_blank">fully homomorphic encryption,</a> or FHE. But there’s a rather large catch. It can take thousands—even tens of thousands—of times as long to compute on today’s CPUs and GPUs than simply working with the decrypted data.</p><p>So universities, startups, and at least one processor giant have been working on specialized chips that could close that gap. Last month at the <a href="https://www.isscc.org/" target="_blank">IEEE International Solid-State Circuits Conference</a> (ISSCC) in San Francisco, <a href="https://www.intel.com/content/www/us/en/homepage.html" target="_blank">Intel</a> demonstrated its answer, Heracles, which sped up FHE computing tasks as much as 5,000-fold compared to a top-of the-line Intel server CPU.</p><p>Startups are racing to beat Intel and each other to commercialization. But <a href="https://www.linkedin.com/in/sanu-mathew-4073742/" target="_blank">Sanu Mathew,</a> who leads security circuits research at Intel, believes the CPU giant has a big lead, because its chip can do more computing than any other FHE accelerator yet built. “Heracles is the first hardware that works at scale,” he says.</p><p>The scale is measurable both physically and in compute performance. While other FHE research chips have been in the range of 10 square millimeters or less, Heracles is about 20 times that size and is built using Intel’s most advanced, 3-nanometer FinFET technology. And it’s flanked inside a liquid-cooled package by two 24-gigabyte <a href="https://spectrum.ieee.org/dram-shortage" target="_blank">high-bandwidth memory </a>chips—a configuration usually seen only in GPUs for training AI.</p><p class="ieee-inbody-related">RELATED: <a href="https://spectrum.ieee.org/how-to-compute-with-data-you-cant-see" target="_blank">How to Compute with Data You Can’t See</a></p><p>In terms of scaling compute performance, Heracles showed muscle in live demonstrations at ISSCC. At its heart the demo was a simple private query to a secure server. It simulated a request by a voter to make sure that her ballot had been registered correctly. The state, in this case, has an encrypted database of voters and their votes. To maintain her privacy, the voter would not want to have her ballot information decrypted at any point; so using FHE, she encrypts her ID and vote and sends it to the government database. There, without decrypting it, the system determines if it is a match and returns an encrypted answer, which she then decrypts on her side.</p><p>On an Intel Xeon server CPU, the process took 15 milliseconds. Heracles did it in 14 microseconds. While that difference isn’t something a single human would notice, verifying 100 million voter ballots adds up to more than 17 days of CPU work versus a mere 23 minutes on Heracles.</p><p>Looking back on the five-year journey to bring the Heracles chip to life, <a href="https://www.linkedin.com/in/ro-cammarota-a226b817/" target="_blank">Ro Cammarota</a>, who led the project at Intel until last December and is now at University of California, Irvine, says “We have proven and delivered everything that we promised.”</p><h2>FHE Data Expansion</h2><p class="rm-anchors" id="fhe">FHE is fundamentally a mathematical transformation, sort of like the Fourier transform. It encrypts data using a quantum-computer-proof algorithm, but, crucially, uses corollaries to the mathematical operations usually used on unencrypted data. These corollaries achieve the same ends on the encrypted data.<strong></strong></p><p>One of the main things holding such secure computing back is the explosion in the size of the data once it’s encrypted for FHE, <a href="https://www.linkedin.com/in/anupamgolder/" target="_blank">Anupam Golder</a>, a research scientist at Intel’s circuits research lab, told engineers at ISSCC. “Usually, the size of cipher text is the same as the size of plain text, but for FHE it’s orders of magnitude larger,” he said.</p><p>While the sheer volume is a big problem, the kinds of computing you need to do with that data is also an issue. FHE is all about very large numbers that must be computed with precision. While a CPU can do that, it’s very slow going—integer addition and multiplication take about 10,000 more clock cycles in FHE. Worse still, CPUs aren’t built to do such computing in parallel. Although GPUs excel at parallel operations, precision is not their strong suit. (In fact, from generation to generation, GPU designers have devoted more and more of the chip’s resources to <a href="https://spectrum.ieee.org/nvidia-gpu" target="_blank">computing less-and-less-precise numbers</a>.)</p><p>FHE also requires some oddball operations with names like “twiddling” and “automorphism,” and it relies on a compute-intensive noise-cancelling process called bootstrapping. None of these things are efficient on a general-purpose processor. So, while clever algorithms and libraries of software cheats have been developed over the years, the need for a hardware accelerator remains if FHE is going to tackle large-scale problems, says Cammarota.</p><h2>The Labors of Heracles</h2><p class="rm-anchors" id="heracles">Heracles was initiated under a <span>Defense Advanced Research Projects Agency</span> (DARPA) program five years ago to accelerate FHE using purpose-built hardware. It was developed as “a whole system-level effort that went all the way from theory and algorithms down to the circuit design,” says Cammarota.</p><p>Among the first problems was how to compute with numbers that were larger than even the 64-bit words that are today a CPU’s most precise. There are ways to break up these gigantic numbers into chunks of bits that can be calculated independently of each other, providing a degree of parallelism. Early on, the Intel team made a big bet that they would be able to make this work in smaller, 32-bit chunks, yet still maintain the needed precision. This decision gave the Heracles architecture some speed and parallelism, because the 32-bit arithmetic circuits are considerably smaller than 64-bit ones, explains Cammarota.</p><p>At Heracles’s heart are 64 compute cores—called tile-pairs—arranged in an eight-by-eight grid. These are what are called single instruction multiple data (SIMD) compute engines designed to do the polynomial math, twiddling, and other things that make up computing in FHE and to do them in parallel. An on-chip 2D mesh network connects the tiles to each other with wide, 512-byte, buses.</p><p class="ieee-inbody-related">RELATED: <a href="https://spectrum.ieee.org/homomorphic-encryption-llm" target="_blank">Tech Keeps Chatbots From Leaking Your Data</a></p><p>Important to making encrypted computing efficient is feeding those huge numbers to the compute cores quickly. The sheer amount of data involved meant linking 48-GB-worth of expensive high-bandwidth memory to the processor with 819-GB-per-second connections. Once on the chip, data musters in 64 megabytes of cache memory—somewhat more than an Nvidia <a href="https://spectrum.ieee.org/nvidias-next-gpu-shows-that-transformers-are-transforming-ai" target="_blank">Hopper-generation GPU</a>. From there it can flow through the array at 9.6 terabytes per second by hopping from tile-pair to tile-pair.</p><p>To ensure that computing and moving data don’t get in each other’s way, Heracles runs three synchronized streams of instructions simultaneously, one for moving data onto and off of the processor, one for moving data within it, and a third for doing the math, Golder explained.</p><p class="rm-anchors" id="faster">It all adds up to some massive speedups, according to Intel. Heracles—operating at 1.2 gigahertz—takes just 39 microseconds to do FHE’s critical math transformation, a 2,355-fold improvement over an Intel Xeon CPU running at 3.5 GHz. Across seven key operations, Heracles was 1,074 to 5,547 times as fast.</p><p>The differing ranges have to do with how much data movement is involved in the operations, explains Mathew. “It’s all about balancing the movement of data with the crunching of numbers,” he says.</p><h2>FHE Competition</h2><p class="rm-anchors" id="commercial">“It’s very good work,” <a href="https://www.linkedin.com/in/kurt-rohloff/" target="_blank">Kurt Rohloff</a>, chief technology officer at FHE software firm <a href="https://dualitytech.com/platform/technology-fully-homomorphic-encryption/" target="_blank">Duality Technology</a>, says of the Heracles results. Duality was part of a team that developed a competing accelerator design under the same DARPA program that brought forth Intel’s Heracles. “When Intel starts talking about scale, that usually carries quite a bit of weight.”</p><p>Duality’s focus is less on new hardware than on software products that do the kind of encrypted queries that Intel demonstrated at ISSCC. At the scale in use today “there’s less of a need for [specialized] hardware,” says Rohloff. “Where you start to need hardware is emerging applications around deeper machine-learning-oriented operations like neural net, LLMs, or semantic search.”</p><p>Last year, Duality demonstrated an <a href="https://spectrum.ieee.org/homomorphic-encryption-llm" target="_self">FHE-encrypted language model called BERT</a>. Like more famous LLMs such as ChatGPT, BERT is a transformer model. However it’s only one-tenth the size of even the most compact LLMs.</p><p><a href="https://www.linkedin.com/in/barrus/" target="_blank">John Barrus</a>, vice president of product at Dayton, Ohio–based <a href="https://niobiummicrosystems.com/" target="_blank">Niobium Microsystems</a>, an FHE chip startup <a href="https://www.galois.com/" target="_blank">spun out</a> of another DARPA competitor, agrees that encrypted AI is a key target of FHE chips. “There are a lot of smaller models that, even with FHE’s data expansion, will run just fine on accelerated hardware,” he says.</p><p>With no stated commercial plans from Intel, Niobium expects its chip to be “the world’s first commercially viable FHE accelerator, designed to enable encrypted computations at speeds practical for real-world cloud and AI infrastructure.” Although it hasn’t announced when a commercial chip will be available, last month the startup revealed that it had inked a deal worth 10 billion South Korean won (US $6.9 million) with Seoul-based chip design firm <a href="https://semifive.com/" target="_blank">Semifive</a> to develop the FHE accelerator for fabrication using Samsung Foundry’s 8-nanometer-process technology.</p><p>Other startups including <a href="https://cornami.com/" target="_blank">Cornami</a>,  <a href="https://www.fabriccryptography.com/" target="_blank">Fabric Cryptography</a>, and <a href="https://optalysys.com/" target="_blank">Optalysys</a> have been working on chips to accelerate FHE. Optalysys CEO <a href="https://optalysys.com/people/" target="_blank">Nick New</a> says Heracles hits about the level of speedup you could hope for using an all-digital system. “We’re looking at pushing way past that digital limit,” he says. His company’s approach is to use the physics of a photonic chip to do FHE’s compute-intensive transform steps. That photonics chip is on its seventh generation, he says, and among the next steps is to 3D integrate it with custom silicon to do the nontransform steps and coordinate the whole process. A full 3D-stacked commercial chip could be ready in two or three years, says New.</p><p>While competitors develop their chips, so will Intel, says Mathew. It will be improving on how much the chip can accelerate computations by fine-tuning the software. It will also be trying out more massive FHE problems, and exploring hardware improvements for a potential next generation. “This is like the first microprocessor…the start of a whole journey,” says Mathew.</p>]]></description><pubDate>Tue, 10 Mar 2026 13:00:04 +0000</pubDate><guid>https://spectrum.ieee.org/fhe-intel</guid><category>Privacy</category><category>Intel</category><category>Encryption</category><category>Homomorphic-encryption</category><category>Hardware-acceleration</category><category>Isscc</category><dc:creator>Samuel K. Moore</dc:creator><media:content medium="image" type="image/jpeg" url="https://spectrum.ieee.org/media-library/overhead-view-of-intel-s-computing-chip-called-heracles.jpg?id=65174073&amp;width=980"></media:content></item><item><title>Finite-Element Approaches to Transformer Harmonic and Transient Analysis</title><link>https://content.knowledgehub.wiley.com/solving-harmonic-and-transient-challenges-in-transformers-using-integrateds-faraday/</link><description><![CDATA[
<img src="https://spectrum.ieee.org/media-library/logo-of-integrated-engineering-software-with-pixelated-geometric-design-and-text.png?id=65106417&width=980"/><br/><br/><p>Explore structured finite-element methodologies for analyzing transformer behavior under harmonic and transient conditions — covering modelling, solver configuration, and result validation techniques.</p><p><strong>What Attendees will Learn</strong><span></span></p><ol><li>How FEM enables pre-fabrication performance evaluation — Assess magnetic field distribution, current behavior, and turns-ratio accuracy through simulation rather than physical testing.</li><li><span>How harmonic analysis uncovers saturation and imbalance — Identify high-flux regions and current asymmetries that analytical methods may not capture.</span></li><li><span>How transient simulations characterize dynamic response — Examine time-domain current waveforms, inrush behavior, and multi-cycle stabilization.</span></li><li><span>How modelling choices affect simulation fidelity — Understand the impact of coil definitions, winding configurations, solver type, and material models on accuracy.</span></li></ol><p><span><a href="https://content.knowledgehub.wiley.com/solving-harmonic-and-transient-challenges-in-transformers-using-integrateds-faraday/" target="_blank">Download this free whitepaper now!</a><br/></span></p>]]></description><pubDate>Tue, 10 Mar 2026 10:00:03 +0000</pubDate><guid>https://content.knowledgehub.wiley.com/solving-harmonic-and-transient-challenges-in-transformers-using-integrateds-faraday/</guid><category>Type-whitepaper</category><category>Transformers</category><category>Finite-element-analysis</category><category>Harmonic</category><dc:creator>Integrated Engineering Software</dc:creator><media:content medium="image" type="image/png" url="https://assets.rbl.ms/65106417/origin.png"></media:content></item></channel></rss>