Cadastre-se agora para um orçamento mais personalizado!

Problem scaling AI? MIT proposes sub-photon optical deep learning at the edge

01 de novembro de 2022 Hi-network.com

MIT scientists used special photodetectors to perform an AI equation just by beaming light to a low-power client device. The approach could be especially useful in deep space. 

Tiernan Ray/

One of the most pressing concerns for the industrial application of artificial intelligence is how to run the programs on small computing devices that have very little processing power, very little memory, and possibly a limit in terms of energy available, in the case of batteries.

Innovation

  • I tried Apple Vision Pro and it's far ahead of where I expected
  • This tiny satellite communicator is packed full of features and peace of mind
  • How to use ChatGPT: Everything you need to know
  • These are my 5 favorite AI tools for work

The so-called edge market for AI has been a huge area of late, with startups receiving tens of millions in venture capital to come up with chips and software. The edge effort has led to special development tools for machine-learning forms of AI, such as the TinyML initiative from Google. 

Those two paths represent two philosophies: Either make edge devices more powerful, or slim down AI programs to use fewer calculations. 

Also: AI's true goal may no longer be intelligence

There is third possible approach, and that is to try and balance more carefully what work is done on constrained devices and by what means. That's the plan put forward in October by MIT researchers in the scholarly journal Science. 

Researcher Alexander Sludds and colleagues at MIT's Research Laboratory of Electronics, Computer Science and Artificial Intelligence Laboratory, and Lincoln Laboratory, in partnership with Nokia and NTT Research, have developed a system that uses photonics to beam data to a client device where it can be calculated in the optical domain in a vastly more energy-efficient manner.

Their network setup, which they call Netcast, can perform the fundamental operation of manipulating the weights, or parameters, of a deep neural network, using about 10 femtoJoules of power, or 10 fJ, which, they relate, "is three orders of magnitude lower than is possible in existing digital CMOS" -- meaning standard semiconductor chips.

A femtoJoule, written as a decimal point followed by 14 zeros and a 1, is one-quadrillionth, with is a very tiny fraction of a joule, a joule being the amount of electricity to run a 1-watt device for a second. 

That tiny, tiny fraction of a watt is a major energy savings and is important because many edge devices, the authors note, will have a total power budget in milliwatts, or thousandths of a watt, versus typical computing devices using tens or hundreds of watts. The femtoJoule operation of Netcast effectively gets the program below what had to date been "a stubborn bottleneck near 1 pJ," aka one picoJoule, or one-trillionth of a joule.

Also: The AI edge chip market is on fire, kindled by 'staggering' VC funding

The key to Netcast was how to reduce the work the client has to perform for the fundamental operation of a neural net in order to get within that 10-femtoJoule budget. 

A neural net makes predictions by passing some input data to its parameters or weights, and multiplying the input by the weight. That mathematical operation, the product of an input vector and a parameter matrix, is called a multiply-accumulate, or MAC, operation, and neural net programs do tons of them every second as the multiple weights of each network layer are applied to the input.

The biggest power hog for most neural nets in general is fetching data from RAM memory chips and accessing the network. That's a problem because the neural weights are usually stored in RAM, so every layer of MAC operations can require multiple trips out over the PCIe bus to RAM and perhaps even to a network line card for remote memory stores. 

Hence, the key to Netcast was how to minimize memory access and network traffic for the client device. 

Sludds et al.

The solution is an existing photonic technology called wavelength division multiplexing. Using WDM, as it's commonly referred to, multiple pieces of data can be sent across a fiber-optic line simultaneously by assigning each piece of data its own wavelength of light so that the multiple data share the total available spectrum of radiation in the fiber. WDM is a very mature, solid technology that is used in all modern telecom networks to increase the capacity of fiber-optic data transmission; it forms the backbone of the Internet. 

Each row of a matrix can be encoded on a wavelength of light and then "broadcast" to the client device, so that a multi-wavelength WDM signal can send an entire weight matrix or even multiple matrices. At the client device, an optical receiver recovers the data encoded in each wavelength and combines it with the input data to perform the matrix multiply in the optical domain rather than electrically. The product can then be stored electrically on local RAM once converted from the optical signal.

Sludds and team write that this results in a dramatic simplification of the components that need to be in a client device at the edge. 

"This architecture minimizes the active components at the client, requiring only a single optical transceiver modulator, digital-to-analog converter (DAC), and analog-to-digital converter (ADC)."

The authors constructed an actual version of Netcast that runs over 84 kilometers of fiber using WDM with a capacity of 2.4 terabits per second, running from the main MIT campus to the Lincoln Lab and back. Their test of the system is to perform predictions on a classic machine learning task, the MNIST database of handwritten characters. The images of the handwritten characters are input to the neural net, and the net has to carry out an image recognition task, identifying which character each picture represents.

"Using 1,000 test images locally, we demonstrate 98.7% accurate computation, comparable with the model's baseline accuracy of 98.7%," they report.

The authors go even further. Anticipating deployment in satellites and other exotic locales, they worked on coming up with photodetectors, called integrating receivers, that can work with very small numbers of photons. 

"Applications of Netcast, including free-space deployment to drones or spacecraft, can operate in deeply photon-starved environments," they write. A version of their integrating receivers is able to detect the results of a MAC operation operating at only fractions of a femtoJoule, known as an attoJoule, which requires only 100 photons for the MAC operation. 

But the authors go even further. They were able to go all the way to a theoretical limit of Netcast where each Mac requires less than a single photon to be detected. Using what are called superconducting nanowire single-photon detectors (SNSPDs), they construct a receiver that can measure the results of each MAC with less than a photon of information. 

Space

  • What is Artemis? Everything you need to know about NASA's new moon mission
  • NASA has solved the mystery of Voyager 1's strange data transmissions
  • NASA's new tiny, high-powered laser could find water on the Moon
  • NASA is blazing an inspirational trail. We need to make sure everyone can follow it

"This result may at first seem surprising given that less than a single photon per MAC is counterintuitive," wrote Sludds and team. "We can understand this measurement better by noting that at readout, we have performed a vector-vector product with M = 100 MACs. Each MAC can have less than a single photon in it, but the measured signal will have many photons in it."

The implications for computing could be profound. 

"The realization of computing with less than one photon per MAC," they wrote, "could enable a new class of computing systems that protect both client input and server weight data" from the standpoint of data privacy. It could also make computing on spacecraft more reliable. "Weight data from a directional base station could be transmitted to the spacecraft and classified on the craft, before the results are transmitted to Earth."

All the parts of Netcast can be made today in any standard semiconductor chip factory, Sludds and team noted.

In concluding, they wrote, "Our approach removes a fundamental bottleneck in edge computing, enabling high-speed computing on deployed sensors and drones."

Artificial Intelligence

The impact of artificial intelligence on software development? Still unclearAndroid 14's AI-generated wallpapers are super fun. Here's how to create themAI aims to predict and fix developer coding errors before disaster strikesGenerative AI is everything, everywhere, all at once
  • The impact of artificial intelligence on software development? Still unclear
  • Android 14's AI-generated wallpapers are super fun. Here's how to create them
  • AI aims to predict and fix developer coding errors before disaster strikes
  • Generative AI is everything, everywhere, all at once

tag-icon Tags quentes : Inteligência artificial Inovação

Copyright © 2014-2024 Hi-Network.com | HAILIAN TECHNOLOGY CO., LIMITED | All Rights Reserved.