Physical learning of power-efficient solutions

Menachem Stern, Sam Dillavou, Dinesh Jayaraman, Douglas J. Durian, Andrea J. Liu
Condensed Matter, Disordered Systems and Neural Networks, Disordered Systems and Neural Networks (cond-mat.dis-nn), Soft Condensed Matter (cond-mat.soft), Statistical Mechanics (cond-mat.stat-mech)
2023-10-15 16:00:00
As the size and ubiquity of artificial intelligence and computational machine learning (ML) models grow, their energy consumption for training and use is rapidly becoming economically and environmentally unsustainable. Neuromorphic computing, or the implementation of ML in hardware, has the potential to reduce this cost. In particular, recent laboratory prototypes of self-learning electronic circuits, examples of ``physical learning machines," open the door to analog hardware that directly employs physics to learn desired functions from examples. In this work, we show that this hardware platform allows for even further reduction of energy consumption by using good initial conditions as well as a new learning algorithm. Using analytical calculations, simulation and experiment, we show that a trade-off emerges when learning dynamics attempt to minimize both the error and the power consumption of the solution--greater power reductions can be achieved at the cost of decreasing solution accuracy. Finally, we demonstrate a practical procedure to weigh the relative importance of error and power minimization, improving power efficiency given a specific tolerance to error.
PDF: Physical learning of power-efficient solutions.pdf
Empowered by ChatGPT