Skip to main content Skip to secondary navigation

Docket #: S18-093

Efficient, Scalable Training of Artificial Neural Networks Directly on Optical Devices

Engineers in Prof. Shanhui Fan's laboratory have developed an efficient, scalable, in-situ method to train, configure and tune complex photonic circuits for artificial intelligence and machine learning. This technology adapts a conventional numerical algorithm (backpropagation) so that it can be implemented experimentally directly on an optical chip. Those physical experiments then provide the input for optimizing the chip toward any designated function. Because photonic circuits are faster, less computationally expensive and more energy efficient than traditional electronic computers, this optics-only in-situ approach is more efficient than conventional training that uses a computational model on an external computer. In addition, the method can be scaled to a large network because it tests and tunes parameters in parallel rather than using brute force serial techniques. This method was originally designed to train artificial neural networks (ANN) in deep learning applications. It could also be utilized to create other complex, adaptive nonlinear photonic circuits with end-user applications such as LiDAR, laser accelerators and optical communications.

Figure description - Researchers have shown a neural network can be trained using an optical circuit (blue rectangle in the illustration). In the full network there would be several of these linked together. The laser inputs (green) encode information that is carried through the chip by optical waveguides (black). The chip performs operations crucial to the artificial neural network using tunable beam splitters, which are represented by the curved sections in the waveguides. These sections couple two adjacent waveguides together and are tuned by adjusting the settings of optical phase shifters (red and blue glowing objects), which act like 'knobs' that can be adjusted during training to perform a given task.

Stage of Research
The inventors have performed optical simulations to demonstrated proof of concept for this training technique by teaching an algorithm to perform complicated functions.

Applications

  • Artificial Neural Network (ANN) training - optimize photonic hardware platforms for artificial intelligence and machine learning with end-user applications such as speech and image recognition
  • Photonic devices - methods could be used for sensitivity analysis of any reconfigurable photonic system with end-user applications such as:
    • LiDAR
    • laser accelerators
    • optical communications
    • scientific research platforms

Advantages

  • Efficient
    • direct optical training realizes computational efficiency, speed and power advantages gained by using photonic circuits instead of traditional solid-state circuits for artificial networks
    • training time reduced because parameters are adjusted in parallel
  • Scalable for large networks
    • training technique adjusts parameters in parallel, unlike brute force gradient methods which operate in serial
    • scales with network size by operating in constant time with respect to the number of parameters
  • Accurate - in-situ training on a physical device avoids model-based methods, which often do not perfectly represent the physical system
  • Can be implemented with no additional hardware or components

Publications

Related Links

Patents

Similar Technologies

Explore similar technologies by keyword: