Breaking the scaling limits of analog computing

MIT researchers have developed a technique that greatly reduces the error in an optical neural network, which uses light to process data instead of electrical signals. With their technique, the larger an optical neural network becomes, the lower the error in its computations. This could enable them to scale these devices up so they would be large enough for commercial uses. Read Entire Article

© 2024 Thiratti. All rights reserved.