image: Postdoctoral fellow Tianwei Wu (left) and Professor Liang Feng (right) in the lab, demonstrating some of the apparatus used to develop the new, light-powered chip.
Credit: Sylvia Zhang.
Penn Engineers have developed the first programmable chip that can train nonlinear neural networks using light — a breakthrough that could dramatically speed up AI training, reduce energy use and even pave the way for fully light-powered computers.
While today’s AI chips are electronic and rely on electricity to perform calculations, the new chip is photonic, meaning it uses beams of light instead. Described in Nature Photonics, the chip reshapes how light behaves to carry out the nonlinear mathematics at the heart of modern AI.
“Nonlinear functions are critical for training deep neural networks,” says Liang Feng, Professor in Materials Science and Engineering (MSE) and in Electrical and Systems Engineering (ESE), and the paper’s senior author. “Our aim was to make this happen in photonics for the first time.”
The Missing Piece in Photonic AI
Most AI systems today depend on neural networks, software designed to mimic biological neural tissue. Just as neurons connect to allow biological creatures to think, neural networks link together layers of simple units, or “nodes,” enabling AI systems to perform complex tasks.
In both artificial and biological systems, these nodes only “fire” once a threshold is reached — a nonlinear process that allows small changes in input to cause larger, more complex changes in output.
Without that nonlinearity, adding layers does nothing: the system just reduces to a single-layer linear operation, where inputs are simply added together, and no real learning occurs.
While many research teams, including teams at Penn Engineering, have developed light-powered chips capable of handling linear mathematical operations, none has solved the challenge of representing nonlinear functions using only light — until now.
“Without nonlinear functions, photonic chips can’t train deep networks or perform truly intelligent tasks,” says Tianwei Wu (Gr’24), a postdoctoral fellow in ESE and the paper’s first author.
Reshaping Light with Light
The team’s breakthrough begins with a special semiconductor material that responds to light. When a beam of "signal" light (carrying the input data) passes through the material, a second "pump" beam shines in from above, adjusting how the material reacts.
By changing the shape and intensity of the pump beam, the team can control how the signal light is absorbed, transmitted or amplified, depending on its intensity and the material’s behavior. This process "programs" the chip to perform different nonlinear functions.
“We’re not changing the chip’s structure,” says Feng. “We’re using light itself to create patterns inside the material, which then reshapes how the light moves through it.”
The result is a reconfigurable system that can express a wide range of mathematical functions depending on the pump pattern. That flexibility allows the chip to learn in real time, adjusting its behavior based on feedback from its output.
Training at the Speed of Light
To test the chip’s potential, the team used the chip to solve benchmark AI problems. The platform achieved over 97% accuracy on a simple nonlinear decision boundary task and over 96% on the well-known Iris flower dataset — a machine learning standard.
In both cases, the photonic chip matched or outperformed traditional digital neural networks, but used fewer operations, and did not need power-hungry electronic components.
In one striking result, just four nonlinear optical connections on the chip were equivalent to 20 linear electronic connections with fixed nonlinear activation functions in a traditional model. That efficiency hints at what’s possible as the architecture scales.
Unlike previous photonic systems — which are fixed after fabrication — the Penn chip starts as a blank canvas. The pump light acts like a brush, drawing reprogrammable instructions into the material.
“This is a true proof-of-concept for a field-programmable photonic computer,” says Feng. “It’s a step toward a future where we can train AI at the speed of light.”
Future Directions
While the current work focuses on polynomials — a flexible family of functions widely used in machine learning — the team believes their approach could enable even more powerful operations in the future, such as exponential or inverse functions. That would pave the way for photonic systems that tackle large-scale tasks like training large language models.
By replacing heat-generating electronics with low-energy optical components, the platform also promises to slash energy consumption in AI data centers — potentially transforming the economics of machine learning.
“This could be the beginning of photonic computing as a serious alternative to electronics,” says Liang. “Penn is the birthplace of ENIAC, the world’s first digital computer — this chip might be the first real step toward a photonic ENIAC.”
This study was conducted at the University of Pennsylvania School of Engineering and Applied Science and supported by the Defense Advanced Research Projects Agency (DARPA) (W911NF-21-1-0340), Office of Naval Research (ONR) (N00014-23-1-2882), and National Science Foundation (NSF) (ECCS-2023780, ECCS-2425529, DRM-2326699, PHYS-1847240, DMR-2326698).
Additional co-authors include Yankun Li of Penn Engineering and Li Ge of the College of Staten Island.
Journal
Nature Photonics
Method of Research
Experimental study
Subject of Research
Not applicable
Article Title
Field-Programmable Photonic Nonlinearity
Article Publication Date
15-Apr-2025
COI Statement
The authors declare no competing interests.