Researchers at the University of California, Davis have developed a nanophotonic-based platform for signal processing and optical computing in algorithm-based neural networks that is faster and more energy-efficient than current technologies.
Current techniques for signal processing and optical computing in algorithm-based neural networks are incredibly energy-intensive. For example, over 90% of the total energy consumed in typical convolutional neural networks occurs during the convolution process itself. In addition, many neuromorphic computing systems are limited to only four direct connections (N-S-E-W), and thus require repeaters to re-transmit their optical signals. Each repeater also consumes additional energy.
Researchers at the University of California, Davis have developed a platform for signal processing and optical computing in neural networks that offers massive parallel information processing. This platform allows for complex functionalities and photonic computing in compact applications for which low signal loss is important. In addition, it reduces overall hardware requirements and allows for increased miniaturization. The technology thus opens applications for multi-layer, convolution neural networks with high quantity processing and low power consumption in handheld devices and other products where equipment size or energy consumption requirements prohibited their use previously.
Nanophotonics, Photonics, Neural Networks, Photonic Integrating Circuit, Personal Devices, Energy Consumption