This is an updated version.
The Godfathers of AI and 2018 ACM Turing Award winners Geoffrey Hinton, Yann LeCun, and Yoshua Bengio shared a stage in New York on Sunday night at an event organized by the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI 2020). The trio of researchers have made deep neural networks a critical component of computing, and in individual talks and a panel discussion they discussed their views on current challenges facing deep learning and where it should be heading.
Introduced in the mid 1980s, deep learning gained traction in the AI community the early 2000s. The year 2012 saw the publication of the CVPR paper Multi-column Deep Neural Networks for Image Classification, which showed how max-pooling CNNs on GPUs could dramatically improve performance on many vision benchmarks; while a similar system introduced months later by Hinton and a University of Toronto team won the large-scale ImageNet competition by a significant margin over shallow machine learning methods. These events are regarded by many as the beginning of a deep learning revolution that has transformed AI.
Deep learning has been applied to speech recognition, image classification, content understanding, self-driving, and much more. And according to LeCun who is now Chief AI Scientist at Facebook the current services offered by Facebook, Instagram, Google, and YouTube are all built around deep learning.
Deep learning does however does have its detractors. Johns Hopkins University Professor and one of the pioneers of computer vision Alan Yuille warned last year that deep learnings potential in computer vision has hit a bottleneck.
We read a lot about the limitations of deep learning today, but most of those are actually limitations of supervised learning, LeCun explained in his talk. Supervised learning typically refers to learning with labelled data. LeCun told the New York audience that unsupervised learning without labels or self-supervised learning as he prefers to call it may be a game changer that ushers in AIs next revolution.
This is an argument that Geoff [Hinton] has been making for decades. I was skeptical for a long time but changed my mind, said LeCun.
There are two approaches to object recognition. Theres the good old-fashioned path based approach, with sensible modular representations, but this typically imposes a lot of hand engineering. And then there are convolutional neural nets (CNNs), which learn everything end to end. CNNs get a huge win by wiring in the fact that if a feature is good in one place, its good somewhere else. But their approach to object recognition is very different from human perception.
This informed the first part of Hintons talk, which he personally directed at LeCun: Its about the problems with CNNs and why theyre rubbish.
CNNs are designed to cope with translations, but theyre not so good at dealing with other effects of changing viewpoints such as rotation and scaling. One obvious approach is to use 4D or 6D maps instead of 2D maps but that is very expensive. And so CNN are typically trained on many different viewpoints in order for them to be able to generalize across viewpoints. Thats not very efficient, Hinton explained. Wed like neural nets to generalize to new viewpoints effortlessly. If it learned to recognize something, then you make it 10 times as big and you rotate it 60 degrees, it shouldnt cause them any problem at all. We know computer graphics is like that and wed like to make neural nets more like that.
Hinton believes the answer is capsules. A capsule is a group of neurons that learns to represent a familiar shape or part. Hinton says the idea is to build more structure into neural networks and hope that the extra structure helps them generalize better. Capsules are an attempt to correct the things that are wrong with CNNs.
The capsules Hinton introduced are Stacked Capsule Auto-encoders, which first appeared at NeurIPS 2019 and are very different in many ways from previous capsule versions from ICLR 2018 and NIPS 2017. These had used discriminative learning. Hinton said even at the time he knew this was a bad idea: I always knew unsupervised learning was the right thing to do so it was bad faith to do the previous models. The 2019 capsules use unsupervised learning.
LeCun noted that although supervised learning has proven successful in for example speech recognition and content understanding, it still requires a large amount of labelled samples. Reinforcement learning works great for games and in simulations, but since it requires too many trials its not really applicable in the real world.
The first challenge LeCun discussed was how models can be expected to learn more with fewer labels, fewer samples or fewer trials.
LeCun now supports the unsupervised learning (self-supervised learning) solution Hinton first proposed some 15 years ago. Basically its the idea of learning to represent the world before learning a task and this is what babies do, LeCun explained, suggesting really figuring out how humans learn so quickly and efficiently may be the key that unlocks self-supervised learnings full potential going forward.
Self-supervised learning is largely responsible for the success of natural language processing (NLP) over the last year and a half or so. The idea is to show a system a piece of text, image, or video input, and train a model to predict the piece thats missing for example to predict missing words in a text, which is what transformers and BERT-like language systems were built to do.
But success of Transformers and BERT et al has not transferred into the image domain because it turns out to be much more difficult to represent uncertainty in prediction on images or in video than it is in text because its not discrete. Its practical to produce distributions over all the words in a dictionary, but its hard to represent distributions over all possible video frames. And this is, in LeCuns view, the main technical problem we have to solve if we want to apply self-supervised learning to a wider variety of modalities like videos.
LeCun proposed one solution may be in latent variable energy-based models: An energy-based model is kind of like a probabilistic model except you dont normalize. And one way to train the energy-based model is to give low energy to samples that you observe and high energy to samples you do not observe.
In his talk, LeCun touched on two other challenges:
LeCun opined that nobody currently seems to have a good answer to either of these two challenges, and said he remains open to and looks forward to any possible ideas.
Yoshua Bengio, meanwhile, has shifted his focus to consciousness. After cognitive neuroscience, he believes the time is ripe for ML to explore consciousness, which he says could bring new priors to help systematic and good generalization. Ultimately, Bengio hopes such a research direction could allow DL to expand from System 1 to System 2 referring to a dichotomy introduced by Daniel Kahneman in his book Thinking, Fast and Slow. System 1 represents what current deep learning is very good at intuitive, fast, automatic, anchored in sensory perception. System 2 meanwhile represents rational, sequential, slow, logical, conscious, and expressible with language.
Before he dived into the valuable lessons that can be learned from consciousness, Bengio briefed the audience on cognitive neuroscience. It used to be seen in the previous century that working on consciousness was kind of taboo in many sciences for all kinds of reasons. But fortunately, this has changed and particularly in cognitive neuroscience. In particular, the Global Workspace Theory by Baars and the recent work in this century based on DeHaene, which really established these theories to explain a lot of the objective neuroscience observations.
Bengio likened conscious processing to a bottleneck and asked Why would this (bottleneck) be meaningful? Why is it that the brain would have this kind of bottleneck where information has to go through this bottleneck, just a few elements to be broadcast to the rest of the brain? Why would we have a short term memory that only contains like six or seven elements? It doesnt make sense.
Bengio said the bottom line is get the magic out of consciousness and proposed the consciousness prior, a new prior for learning representations of high-level concepts of the kind human beings manipulate with language. The consciousness prior is inspired by cognitive neuroscience theories of consciousness. This prior can be combined with other priors in order to help in disentangling abstract factors from each other. What this is saying is that at that level of representation, our knowledge is represented in this very sparse graph where each of the dependencies, these factors involve two, three, four or five entities and thats it.
Consciousness can also provide inspiration on how to build models. Bengio explained Agents are at the particular time at a particular place and they do something and they have an effect. And eventually that effect could have constant consequences all over the universe, but it takes time. And so if we can build models of the world where we have the right abstractions, where we can pin down those changes to just one or a few variables, then we will be able to adapt to those changes because we dont need as much data, as much observation in order to figure out what has changed.
So whats required if deep learning is going to reach human-level intelligence? Bengio referenced his previous suggestions, that missing pieces of the puzzle include:
In a panel discussion, Hinton, LeCun and Bengio were asked how they reconcile their research approaches with colleagues committed to more traditional methods. Hinton had been conspicuously absent from some AAAI conferences, and hinted at why in responding: The last time I submitted a paper to AAAI, I got the worst review I ever got. And it was mean. It said Hinton has been working on this idea for seven years [vector representations] and nobodys interested. Time to move on.
Hinton spoke of his efforts to find a common ground and move on: Right now were in a position where we should just say, lets forget the past and lets see if we can take the idea of doing gradient descent in great big system parameters. And lets see if we can take that idea, because thats really all weve discovered so far. That really works. The fact that that works is amazing. And lets see if we can learn to do reasoning like that.
Author: Fangyu Cai & Yuan Yuan | Editor: Michael Sarazen
Like Loading...
Read more:
AAAI 2020 | Whats Next for Deep Learning? Hinton, LeCun, and Bengio Share Their Visions - Synced
- New neuroscience research sheds light on distinct patterns of learning and generalization in autistic adults - PsyPost - January 23rd, 2025 [January 23rd, 2025]
- Neuroscientists need to do better at explaining basic mental health research - The Transmitter: Neuroscience News and Perspectives - January 23rd, 2025 [January 23rd, 2025]
- How Severance shows the possibilities of cognitive neuroscience - Fast Company - January 23rd, 2025 [January 23rd, 2025]
- AdventHealth Welcomes New Leadership In Heart and Vascular Services, Neuroscience and Orthopedics - Northwest Georgia News - January 23rd, 2025 [January 23rd, 2025]
- School of Neuroscience and Language Sciences Program recognized with University Exemplary Department or Program Award - Virginia Tech - January 23rd, 2025 [January 23rd, 2025]
- Early Exposure to Violent Media Linked to Teen Antisocial Behavior - Neuroscience News - January 23rd, 2025 [January 23rd, 2025]
- The Real Cognitive Neuroscience Behind Severance - WIRED - January 23rd, 2025 [January 23rd, 2025]
- The 15 most popular psychology and neuroscience studies in 2024 - PsyPost - January 1st, 2025 [January 1st, 2025]
- The 'lizard brain' lie: How neuroscience demolished the greatest mind myth - BBC Science Focus - January 1st, 2025 [January 1st, 2025]
- Revolutionizing Brain Diagnostics with Light and AI - Neuroscience News - January 1st, 2025 [January 1st, 2025]
- How Early Experiences Shape Genes, Brain Health, and Resilience - Neuroscience News - January 1st, 2025 [January 1st, 2025]
- A nation exhausted: The neuroscience of why Americans are tuning out political news - Indiana Capital Chronicle - January 1st, 2025 [January 1st, 2025]
- Lithium Restores Brain Function and Behavior in Autism - Neuroscience News - January 1st, 2025 [January 1st, 2025]
- Partners in Diversity presents the science of belonging: exploring the neuroscience of inclusion - Here is Oregon - January 1st, 2025 [January 1st, 2025]
- Classical vs. Operant Conditioning: The Brain's Memory Tug-of-War - Neuroscience News - January 1st, 2025 [January 1st, 2025]
- The Personality Gap Between Singles and the Partnered - Neuroscience News - January 1st, 2025 [January 1st, 2025]
- The Neuroscience Behind Vermeers Girl and Its Hypnotic Power - ZME Science - January 1st, 2025 [January 1st, 2025]
- Serotonin, GABA, and Dopamine Drive Hunger and Feeding - Neuroscience News - December 23rd, 2024 [December 23rd, 2024]
- A nation exhausted: The neuroscience of why Americans are tuning out politics - The Conversation - December 23rd, 2024 [December 23rd, 2024]
- UNO Goalie and Neuroscience Grad Shines in Her Athletic and Academic Aspirations - University of Nebraska Omaha - December 23rd, 2024 [December 23rd, 2024]
- Neuroscience Major Seeks to Bridge the Generation Gap, Help Alzheimers Patients - Pomona College - December 23rd, 2024 [December 23rd, 2024]
- Spectrum 2024: Year in review - The Transmitter: Neuroscience News and Perspectives - December 23rd, 2024 [December 23rd, 2024]
- Say what? The Transmitters top quotes of 2024 - The Transmitter: Neuroscience News and Perspectives - December 23rd, 2024 [December 23rd, 2024]
- Targeted or Broadcast? How the Brain Processes Visual Information - Neuroscience News - December 23rd, 2024 [December 23rd, 2024]
- 70 Is the New 60: Age Related Declines Slowing in Older People - Neuroscience News - December 23rd, 2024 [December 23rd, 2024]
- Breathing Rhythms During Sleep Strengthen Memory Consolidation - Neuroscience News - December 23rd, 2024 [December 23rd, 2024]
- How our brains think: Exploring the world of neuroscience at the Yale Peabody Museum - Connecticut Public - December 23rd, 2024 [December 23rd, 2024]
- Assembloids illuminate circuit-level changes linked to autism, neurodevelopment - The Transmitter: Neuroscience News and Perspectives - December 23rd, 2024 [December 23rd, 2024]
- Mapping the Brain's Response to Social Rejection - Neuroscience News - December 9th, 2024 [December 9th, 2024]
- An eye for science: Q&A with Bryan W. Jones - The Transmitter: Neuroscience News and Perspectives - December 9th, 2024 [December 9th, 2024]
- Short Sleep and High Blood Pressure Linked to Brain Aging - Neuroscience News - December 9th, 2024 [December 9th, 2024]
- Neighborhood Disadvantage Linked to Cognitive Health Risks - Neuroscience News - December 9th, 2024 [December 9th, 2024]
- Psychosis Risk Tied to Heavy Cannabis Use and Genetic Factors - Neuroscience News - December 9th, 2024 [December 9th, 2024]
- Most Teens Recover From Long Covid Within Two Years - Neuroscience News - December 9th, 2024 [December 9th, 2024]
- Opportunities and challenges of single-cell and spatially resolved genomics methods for neuroscience discovery - Nature.com - December 9th, 2024 [December 9th, 2024]
- How Evolution Shaped the Brains Understanding of Numbers - Neuroscience News - December 9th, 2024 [December 9th, 2024]
- Neuroscience Study Aboard Cunard's Queen Mary 2 Reveals Cognitive Benefits of Slow Travel at Sea - PR Newswire - November 28th, 2024 [November 28th, 2024]
- How Expectations Shape Our Gaze in a Changing World - Neuroscience News - November 28th, 2024 [November 28th, 2024]
- To keep or not to keep: Neurophysiologys data dilemma - The Transmitter: Neuroscience News and Perspectives - November 28th, 2024 [November 28th, 2024]
- Does Alcohol Consumption Contribute to Hair Loss? - Neuroscience News - November 28th, 2024 [November 28th, 2024]
- Brains Traffic Controllers Hold Key to Learning and Memory - Neuroscience News - November 28th, 2024 [November 28th, 2024]
- Despite Neuroscience Setback, AbbVie Has Strong Recovery Ahead (ABBV) - Seeking Alpha - November 28th, 2024 [November 28th, 2024]
- Neuroscientists reeling from past cuts advocate for more BRAIN Initiative funding - The Transmitter: Neuroscience News and Perspectives - November 28th, 2024 [November 28th, 2024]
- Want Better Habits? Neuroscience Says This Is How to Train Your Brain - Inc. - November 28th, 2024 [November 28th, 2024]
- Dopamine and Serotonin Work in Opposition for Effective Learning - Neuroscience News - November 28th, 2024 [November 28th, 2024]
- Cunard Proves the Healing Power of Ocean Travel with Breakthrough Neuroscience Research - Travel And Tour World - November 28th, 2024 [November 28th, 2024]
- Bridging the Gap between Meditation, Neuroscience, and the Soul - openPR - November 28th, 2024 [November 28th, 2024]
- Animal Characters in Childrens Books Boost Theory of Mind - Neuroscience News - November 28th, 2024 [November 28th, 2024]
- Emotional Struggles and Tantrums in Preschoolers Linked to ADHD - Neuroscience News - November 28th, 2024 [November 28th, 2024]
- Neuroscience Says This Simple Habit Improves Cognitive Health and Makes Your Brain Act Younger - Inc. - November 20th, 2024 [November 20th, 2024]
- Premature declarations on animal consciousness hinder progress - The Transmitter: Neuroscience News and Perspectives - November 20th, 2024 [November 20th, 2024]
- Medtronic Q2 Earnings: Diabetes And Neuroscience Revenue Boost Growth, Raises Annual Outlook - Yahoo Finance - November 20th, 2024 [November 20th, 2024]
- Trace Neuroscience Nets $101M in Series A Funding for ALS, Dementia Therapy Development - Senior Housing News - November 20th, 2024 [November 20th, 2024]
- How to be a multidisciplinary neuroscientist - The Transmitter: Neuroscience News and Perspectives - November 20th, 2024 [November 20th, 2024]
- Neuroscience Market Expected to Reach USD 71.0 Billion by - GlobeNewswire - November 20th, 2024 [November 20th, 2024]
- Finger-Prick Test Brings Alzheimers Detection Closer to Everyone - Neuroscience News - November 20th, 2024 [November 20th, 2024]
- Dual-Gene Therapy Shows Promise for Hearing and Vision Loss - Neuroscience News - November 20th, 2024 [November 20th, 2024]
- Robots Help Unlock the Mystery of Human Sense of Self - Neuroscience News - November 20th, 2024 [November 20th, 2024]
- The neuroscience of sleep - University of South Carolina - November 20th, 2024 [November 20th, 2024]
- Stress warps fear memories in multiple ways - The Transmitter: Neuroscience News and Perspectives - November 20th, 2024 [November 20th, 2024]
- Mental Exhaustion Drives Aggressive Behavior - Neuroscience News - November 12th, 2024 [November 12th, 2024]
- NeuroAI: A field born from the symbiosis between neuroscience, AI - The Transmitter: Neuroscience News and Perspectives - November 12th, 2024 [November 12th, 2024]
- The neuroscience of deeper learning in math - SmartBrief - November 12th, 2024 [November 12th, 2024]
- What the brain can teach artificial neural networks - The Transmitter: Neuroscience News and Perspectives - November 12th, 2024 [November 12th, 2024]
- How Anthony Zador thinks neuroscience can help improve AI - The Transmitter: Neuroscience News and Perspectives - November 12th, 2024 [November 12th, 2024]
- Discovering Cancer Therapies through Neuroscience - The New York Academy of Sciences - November 12th, 2024 [November 12th, 2024]
- Neuroscience Market Projected to Reach USD 50.2 Billion by 2032, Growing at a 4.0% CAGR S&S Insider - GlobeNewswire - November 12th, 2024 [November 12th, 2024]
- Insights on Brain Aging and Lifelong Cognitive Health - Neuroscience News - November 12th, 2024 [November 12th, 2024]
- A neuroscience PhD student at the University of Oxford has died - The Tab - November 12th, 2024 [November 12th, 2024]
- Exploring the connection between autism and sleep - The Transmitter: Neuroscience News and Perspectives - November 12th, 2024 [November 12th, 2024]
- Astrocytes star in memory storage, recall - The Transmitter: Neuroscience News and Perspectives - November 12th, 2024 [November 12th, 2024]
- Gut Bacteria Modulate Stress Responses Over Time - Neuroscience News - November 12th, 2024 [November 12th, 2024]
- Gut Bacteria Could Hold the Key to Promoting Healthy Aging - Neuroscience News - November 12th, 2024 [November 12th, 2024]
- Microglias pruning function called into question - The Transmitter: Neuroscience News and Perspectives - October 26th, 2024 [October 26th, 2024]
- Depression Alters Brain Circuits, Heightening Negative Perception - Neuroscience News - October 26th, 2024 [October 26th, 2024]
- UNE Researchers Showcase Groundbreaking Work at Global Neuroscience Conference - University of New England - October 26th, 2024 [October 26th, 2024]
- Scientists discover "glue" that holds memory together in fascinating neuroscience breakthrough - PsyPost - October 26th, 2024 [October 26th, 2024]
- Systems neuroscience: combining theory and neurotechnology for a multiscale account of the brain - Nature.com - October 26th, 2024 [October 26th, 2024]
- Seaport Therapeutics adds another $225 million to coffers to embrace the golden age of neuroscience - STAT - October 26th, 2024 [October 26th, 2024]
- ANRO Investors Have Opportunity to Join Alto Neuroscience, Inc. Fraud Investigation with the Schall Law Firm - Business Wire - October 26th, 2024 [October 26th, 2024]