July 10th, 2024

Training of Physical Neural Networks

Physical Neural Networks (PNNs) leverage physical systems for computation, offering potential in AI. Research explores training larger models for local inference on edge devices. Various training methods are investigated, aiming to revolutionize AI systems by considering hardware physics constraints.

Read original articleLink Icon
Training of Physical Neural Networks

The paper discusses Physical Neural Networks (PNNs), which utilize the properties of physical systems for computation, presenting a potential opportunity in modern AI. The authors explore the possibility of training AI models significantly larger than current ones and enabling local and private inference on edge devices like smartphones. Various training methods for PNNs are being investigated, including backpropagation-based and backpropagation-free approaches. While no method has yet matched the scale and performance of the widely used backpropagation algorithm in deep learning, the field is evolving rapidly. The research suggests that PNNs could revolutionize AI systems by creating more efficient current-scale models and enabling unprecedented-scale models. The study emphasizes the importance of considering hardware physics constraints in rethinking AI model functioning and training methods. This work highlights the potential for PNNs to reshape the landscape of AI systems with further research and development.

Related

Researchers upend AI status quo by eliminating matrix multiplication in LLMs

Researchers upend AI status quo by eliminating matrix multiplication in LLMs

Researchers innovate AI language models by eliminating matrix multiplication, enhancing efficiency. A MatMul-free method reduces power consumption, costs, and challenges the necessity of matrix multiplication in high-performing models.

An Analog Network of Resistors Promises Machine Learning Without a Processor

An Analog Network of Resistors Promises Machine Learning Without a Processor

Researchers at the University of Pennsylvania created an analog resistor network for machine learning, offering energy efficiency and enhanced computational capabilities. The network, supervised by Arduino Due, shows promise in diverse tasks.

Analysing 16,625 papers to figure out where AI is headed next (2019)

Analysing 16,625 papers to figure out where AI is headed next (2019)

MIT Technology Review analyzed 16,625 AI papers, noting deep learning's potential decline. Trends include shifts to machine learning, neural networks' rise, and reinforcement learning growth. AI techniques cycle, with future dominance uncertain.

Open and remotely accessible Neuroplatform for research in wetware computing

Open and remotely accessible Neuroplatform for research in wetware computing

An open Neuroplatform for wetware computing research combines electrophysiology and AI with living neurons. It enables long-term experiments on brain organoids remotely, supporting complex studies for energy-efficient computing advancements.

From GPT-4 to AGI: Counting the OOMs

From GPT-4 to AGI: Counting the OOMs

The article discusses AI advancements from GPT-2 to GPT-4, highlighting progress towards Artificial General Intelligence by 2027. It emphasizes model improvements, automation potential, and the need for awareness in AI development.

Link Icon 9 comments
By @tomxor - 3 months
Last time I read about this the main practical difficulty was model transferability.

The very thing that makes it so powerful and efficient is also the thing that make it uncopiable, because sensitivity to tiny physical differences in the devices inevitably gets encoded into the model during training.

It seems intuitive this is an unavoidable, fundamental problem. Maybe that scares away big tech, but I quite like the idea of having invaluable, non-transferable, irreplaceable little devices. Not so easily deprecated by technological advances, flying in the face of consumerism, getting better with age, making people want to hold onto things.

By @ksd482 - 3 months
PNNs resemble neural networks, however at least part of the system is analog rather than digital, meaning that part or all the input/output data is encoded continuously in a physical parameter, and the weights can also be physical, with the ultimate goal of surpassing digital hardware in performance or efficiency.

I am trying to understand what format does a node take in PNNs. Is it a transistor? Or is it more complex than that? Or, is it a combination of a few things such as analog signal and some other sensors which work together to form a single node that looks like the one we are all familiar with?

Can anyone please help me understand what exactly is "physical" about PNNs?

By @Shawnecy - 3 months
My knowledge in this area is incredibly limited, but I figured the paper would mention NanoWire Networks (NWNs) as an emerging physical neural network[0].

Last year, researchers from the University of Sydney and UCLA used NWNs to demonstrate online learning of handwritten digits with an accuracy of 93%.

[0] = https://www.nature.com/articles/s41467-023-42470-5

By @programjames - 3 months
> These methods are typically slow because the number of gradient updates scales linearly with the number of learnable parameters in the network, posing a significant challenge for scaling up.

This is a pretty big problem, though if you use information-bottleneck training you can train each layer simultaneously.

By @UncleOxidant - 3 months
So it sounds like these PNNs are essentially analog implementations of neural nets? Seems like an odd choice of naming to call them 'physical'.
By @craigmart - 3 months
Schools?