News: 0158064971

  ARM Give a man a fire and he's warm for a day, but set fire to him and he's warm for the rest of his life (Terry Pratchett, Jingo)

Researchers Build AI That Builds AI (quantamagazine.org)

(Wednesday January 26, 2022 @05:50PM (msmash) from the AIception dept.)


By using hypernetworks, researchers can now preemptively fine-tune artificial neural networks, [1]saving some of the time and expense of training . From a report:

> Artificial intelligence is largely a numbers game. When deep neural networks, a form of AI that learns to discern patterns in data, began surpassing traditional algorithms 10 years ago, it was because we finally had enough data and processing power to make full use of them. Today's neural networks are even hungrier for data and power. Training them requires carefully tuning the values of millions or even billions of parameters that characterize these networks, representing the strengths of the connections between artificial neurons. The goal is to find nearly ideal values for them, a process known as optimization, but training the networks to reach this point isn't easy. "Training could take days, weeks or even months," said Petar Velickovic, a staff research scientist at DeepMind in London. That may soon change.

>

> Boris Knyazev of the University of Guelph in Ontario and his colleagues have designed and trained a "hypernetwork" -- a kind of overlord of other neural networks -- that could speed up the training process. Given a new, untrained deep neural network designed for some task, the hypernetwork predicts the parameters for the new network in fractions of a second, and in theory could make training unnecessary. Because the hypernetwork learns the extremely complex patterns in the designs of deep neural networks, the work may also have deeper theoretical implications. For now, the hypernetwork performs surprisingly well in certain settings, but there's still room for it to grow -- which is only natural given the magnitude of the problem. If they can solve it, "this will be pretty impactful across the board for machine learning," said Velickovic.



[1] https://www.quantamagazine.org/researchers-build-ai-that-builds-ai-20220125/



Tuning? (Score:3)

by jythie ( 914043 )

So.. trying to make sense of the actual piece.. initially it sounded like they were talking about just architecture or hyperparameters, but later it sounds like it is also pre-populating the weights since they were comparing it to things like resnet? I am really unclear. Though I am terribly amused that they seem worried about a black box creating a black box with no way to explain the mistakes.. the horse is kinda out of the barn on that one.

Re:Tuning? (Score:4, Interesting)

by Thelasko ( 1196535 )

Yeah, my first thought is they were tuning [1]hyperparameters [wikipedia.org] too. (Hyperparameters are settings that tell the training algorithm how to search for coefficients) My first time tuning an ANN I thought it was kind of obvious to do so. I wound up creating a simple search algorithm to calibrate those parameters. The whole hyperparameter tuning process is kinda meta.

Most machine learning articles use so much jargon, it's difficult to tell what's a breakthrough and what's bullshit.

[1] https://en.wikipedia.org/wiki/Hyperparameter_(machine_learning)

Watching It All Unfold... (Score:2)

by IonOtter ( 629215 )

Hmmmm.

"Hungry for data and power", "carefully tuning values" and "overlord".

Yeah, see, those are three phrases and/or concepts that have absolutely no place in the context of artificial intelligence.

It's almost as if they're taking their cues FROM Hollywood.

Machines making machines... (Score:3)

by garyisabusyguy ( 732330 )

what could go wrong with that?

Just remember the Real Rule of Robotics, "Humans are fascinating and we want to make sure they are having a good time"

"performs surprisingly well in certain settings" (Score:2)

by fleeped ( 1945926 )

aka one-trick pony (impressive sure), with the potential to learn another trick or two, for the next round of press release and funding/grant applications

Re: (Score:2)

by gweihir ( 88907 )

Pretty much yes. A lot of research has degraded from actually being useful to giving the appearance of being useful.

But Do They... (Score:2)

by crunchygranola ( 1954152 )

> Researchers Build AI That Builds AI

But do they have an AI to build that ?

Not just "enough data and processing power" (Score:2)

by alispguru ( 72689 )

Modern machine learning benefitted from two recent advances:

* GPU-based neural network processing

* Faster learning algorithms, so networks could converge in a reasonable time

See [1]here [machinelea...owledge.ai] for a historical timeline.

GPU use started around 2008.

Theoretical work that sped up learning in 2006 and 2011.

[1] https://machinelearningknowledge.ai/brief-history-of-deep-learning/

Re: (Score:2)

by raftpeople ( 844215 )

I would probably rephrase "faster learning algorithm" to "a deep network training algorithm", meaning the discovery of the algorithm (by Hinton and group) opened the floodgates.

Prior to that there wasn't really any good algorithm for this problem, other than evolving the network with gradient descent type techniques (sure that is technically an algorithm, but more of a brute force thing of trying random variations vs an actual update formula)

Yo dawg (Score:2)

by OrangeTide ( 124937 )

I heard you liked AI so I added AI to your AI.

At least we will soon be able to put the crackpot theories of paranoid futurists to the test. Of course I hope those guys were all crackpots ...

Archer: "Do you want Skynet?" (Score:2)

by UnknowingFool ( 672806 )

This is how you get Skynet.

Re: (Score:1)

by tom_asdf ( 8560347 )

I agree. Just wait for this "AI" to get into the military computers and Skynet will take over the world.

Deja Vu (Score:1)

by msobel ( 661289 )

It oversees other AI, why don't we call it Skynet?

Re: (Score:2)

by Impy the Impiuos Imp ( 442658 )

Westworld is a closer approximation. Robots designed better robots. Several generations later, humans no longer understood how they worked.

Think how hard it is to detail how biological life works. Now imagine how intricate and specialized and interleaved sections of a giant AI brain might get.

Machines building machines. How perverse. (Score:2)

by mmell ( 832646 )

Shut me down!

When the great Tao is forgotten,
Kindness and morality arise.
When wisdom and intelligence are born,
The great pretense begins.

When there is no peace within the family,
Filial piety and devotion arise.
When the country is confused and in chaos,
Loyal ministers appear.