MITB Banner

Can Neural Networks Function Without Learning Any Weight Parameters?

Share

An Antarctic eelpout swims gracefully in cold dark depths without freezing its internal juices. It does this with the help of anti-freezing proteins (AFPs) which are by-products of million mutations over millions of years of evolution. The 3D structure of AFPs allows them to bind to ice crystals and prevent organisms from freezing by forming a hydrophobic layer that separates liquid from crystallising. This information of proteins is written in genes and encoded in DNA.

Many traits are inherent to an organism. Ducklings swim and the newly hatched chicks try to dodge predators. These species are precocial. They don’t need pre-training. They build upon these skills by adjusting to the environment and rarely suffer from a starting problem.

But when an attempt is made to mimic this naturally occurring intelligence through algorithms, many limitations surface. There are factors like artificially induced biases and the inability of the algorithms to capitalise on randomness.

In an attempt to address this, reinforcement learning has been introduced many years ago where the agent learns to perform tasks better through a reward system. The room for improvement in the case of reinforcement learning algorithms is obviously huge.

Acknowledging the shortcomings of the existing networks, in their paper titled, Weight Agnostic Neural Networks(WANNs), the researchers at Google Brain, dive deep to tweak the functioning of the networks at the foundational level.

In this work, the authors try to de-emphasize the importance of weights and how models can be made to learn better.

They recommend the following procedure:

  1. Create an initial population of minimal neural network topologies.
  2. Evaluate each network over multiple rollouts, with a different shared weight value assigned at each rollout.
  3. Rank networks according to their performance and complexity.
  4. Create a new population by varying the highest ranked network topologies, chosen probabilistically through tournament selection.
  5. The algorithm then repeats from (2), yielding weight agnostic topologies of gradually increasing complexity that perform better over successive generations.
Source: Google Brain

WANNs have experimented on control tasks such as CartPoleSwingUp, Bipedal Walker and Car Racing.

For example, in the case of Bipedal Walker, the task is to guide a two-legged agent across randomly generated terrain. Rewards are awarded for distance travelled, with a cost for motor torque to encourage efficient movement. Each leg is controlled by a hip and knee joint in reaction to 24 inputs, including LIDAR sensors which detect the terrain and proprioceptive information such as the agent’s joint speeds.

Weight Agnostic Neural Networks(WANNs) were evaluated for performance based on 4 conditions:

  • Random weights:  individual weights drawn from range (-2,2)
  • Random shared weight:  a single shared weight drawn from range (-2,2)
  • Tuned shared weight:  the highest performing shared weight value in range.
  • Tuned weights:  individual weights tuned using population-based REINFORCE.

Key Takeaways

Computers, which are super-rational, are expected to show dedication towards a job without any prejudice. But the data provided is generated by humans, curated by humans and then there is this ambitious pursuit towards AGI(human-like intelligence). This is a walk on a tightrope for AI researchers who have to train the models for sophistication on a human level while being pragmatic about the flaws.

Here are a few outcomes of the WANNs:

  • A serious attempt to develop networks that can be described with a minimal description length.
  • WANNs which encode relationships between inputs are well suited to RL tasks: low-dimensional inputs coupled with internal states and environmental interaction allow discovery of reactive and adaptive controllers.
  • Even in the high-dimensional classification task, WANNs perform remarkably well.
  • This ability to quickly fine-tune weights is useful in few-shot learning.
  • It is possible to learn a pruned network capable of performing additional tasks without learning weights

“We hope this work will encourage further research that facilitates the discovery of new architectures that not only possess inductive biases for practical domains but can also be trained with algorithms that may not require gradient computation,” wrote the authors.

Read more about the work here.

Share
Picture of Ram Sagar

Ram Sagar

I have a master's degree in Robotics and I write about machine learning advancements.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.