Intel’s Silicon Photonics Work Could Supercharge AI Neural Networks
The various types of neural networks companies have experimented with so far have all been silicon-based. CPUs, GPUs, TPUs, and FPGAs may have different performance characteristics, but they’re built from the same sorts of materials. Intel, however, is working on building AI networks using silicon photonics, defined as the study and application of photonic systems that use silicon as an optical medium.
Two years ago, research work from MIT demonstrated that optical neural networks (ONNs) could be extremely useful for low-power, low-latency operations. This is possible because a common element of photonic circuits, known as a Mach-Zehnder interferometer (MZI), can be configured to perform 2×2 matrix multiplication. Placing the MZI’s in a triangular mesh allows for larger matrices to be created.
Casimir Wierzynski, Senior Director, Office of the CTO, Artificial Intelligence Products Group at Intel, writes:
As in any manufacturing process, there are imperfections, which means that there will be small variations within and across chips, and these will affect the accuracy of computations. In order to move ONNs closer to production, we wanted to understand how sensitive they were to typical process variations, especially as they scaled up to more realistic problem sizes. We also wanted to know whether we could make them more robust to these variations by considering different circuit architectures.
Intel’s new paper is a survey of two types of fault-resistant ONN. One of them had a more tunable design (GridNet), while the other was built with better fault-tolerance (FFTNet). As the name implies, Gridnet is designed as a grid, while FFTNet “arranges the MZIs in a butterfly-like pattern modeled after architectures for computing Fast Fourier Transforms.”
Both ONNs were then trained to recognize handwriting. GridNet proved more accurate than FFTNet, at 98 percent to 95 percent, but FFTNet was specifically more robust when it came to handling manufacturing imprecision, which was simulated by adding noise and phase-shifting to each MZI. After simulating the real-world impact of this noise, GridNet became just 50 percent effective at handwriting recognition. FFTNet’s recognition level remained nearly constant.
These simulations and prototype work suggest that optical neural networks could be valid alternatives to silicon-based designs. Wierzynski writes:
Larger circuits will require more devices, such as MZIs, per chip. Therefore, attempting to “fine tune” each device on a chip post-manufacturing will be a growing challenge. A more scalable strategy will be to train ONNs in software, then mass produce circuits based on those parameters. Our results suggest that choosing the right architecture in advance can greatly increase the probability that the resulting circuits will achieve their desired performance even in the face of manufacturing variations.
Being able to build effective ONNs even in the face of manufacturing variations means it is much easier to build them effectively while learning how to build them in the first place. That kind of capability could help with commercialization if optical architectures can scale up and compete with conventional silicon work.
Continue reading
New Photonic Microchip Mimics Basic Brain Function
Our current computational architecture can't match the efficiency of our own minds, but researchers have discovered a way to begin narrowing that gap by creating a small artificial neurosynaptic network powered by light.