ScienceDaily
Your source for the latest research news
Follow Facebook Twitter LinkedIn Subscribe RSS Feeds Newsletters
New:
  • Alligators Can Regrow Their Tails Too
  • Two Liquid States of Water Exist
  • Zebra Finches Unmask the Bird Behind the Song
  • Most Effective Strategies to Cut COVID-19 Spread
  • Memory 'Fingerprints' Reveal Brain Organization
  • A Biochemical Random Number
  • Geology at Mars' Equator: Ancient Megaflood
  • How the Brain Forms Sensory Memories
  • Healthy Sleep Habits Cut Risk of Heart Failure
  • NASA's SpaceX Crew-1 Astronauts Headed to ISS
advertisement
Follow all of ScienceDaily's latest research news and top science headlines!
Science News
from research organizations

1

2

Shrinking massive neural networks used to model language

Date:
December 1, 2020
Source:
Massachusetts Institute of Technology
Summary:
Deep learning neural networks can be massive, demanding major computing power. In a test of the 'lottery ticket hypothesis,' researchers have found leaner, more efficient subnetworks hidden within BERT models. The discovery could make natural language processing more accessible.
Share:
FULL STORY

You don't need a sledgehammer to crack a nut.

advertisement

Jonathan Frankle is researching artificial intelligence -- not noshing pistachios -- but the same philosophy applies to his "lottery ticket hypothesis." It posits that, hidden within massive neural networks, leaner subnetworks can complete the same task more efficiently. The trick is finding those "lucky" subnetworks, dubbed winning lottery tickets.

In a new paper, Frankle and colleagues discovered such subnetworks lurking within BERT, a state-of-the-art neural network approach to natural language processing (NLP). As a branch of artificial intelligence, NLP aims to decipher and analyze human language, with applications like predictive text generation or online chatbots. In computational terms, BERT is bulky, typically demanding supercomputing power unavailable to most users. Access to BERT's winning lottery ticket could level the playing field, potentially allowing more users to develop effective NLP tools on a smartphone -- no sledgehammer needed.

"We're hitting the point where we're going to have to make these models leaner and more efficient," says Frankle, adding that this advance could one day "reduce barriers to entry" for NLP.

Frankle, a PhD student in Michael Carbin's group at the MIT Computer Science and Artificial Intelligence Laboratory, co-authored the study, which will be presented next month at the Conference on Neural Information Processing Systems. Tianlong Chen of the University of Texas at Austin is the lead author of the paper, which included collaborators Zhangyang Wang, also of Texas A&M, as well as Shiyu Chang, Sijia Liu, and Yang Zhang, all of the MIT-IBM Watson AI Lab.

You've probably interacted with a BERT network today. It's one of the technologies that underlies Google's search engine, and it has sparked excitement among researchers since Google released BERT in 2018. BERT is a method of creating neural networks -- algorithms that use layered nodes, or "neurons," to learn to perform a task through training on numerous examples. BERT is trained by repeatedly attempting to fill in words left out of a passage of writing, and its power lies in the gargantuan size of this initial training dataset. Users can then fine-tune BERT's neural network to a particular task, like building a customer-service chatbot. But wrangling BERT takes a ton of processing power.

advertisement

"A standard BERT model these days -- the garden variety -- has 340 million parameters," says Frankle, adding that the number can reach 1 billion. Fine-tuning such a massive network can require a supercomputer. "This is just obscenely expensive. This is way beyond the computing capability of you or me."

Chen agrees. Despite BERT's burst in popularity, such models "suffer from enormous network size," he says. Luckily, "the lottery ticket hypothesis seems to be a solution."

To cut computing costs, Chen and colleagues sought to pinpoint a smaller model concealed within BERT. They experimented by iteratively pruning parameters from the full BERT network, then comparing the new subnetwork's performance to that of the original BERT model. They ran this comparison for a range of NLP tasks, from answering questions to filling the blank word in a sentence.

The researchers found successful subnetworks that were 40 to 90 percent slimmer than the initial BERT model, depending on the task. Plus, they were able to identify those winning lottery tickets before running any task-specific fine-tuning -- a finding that could further minimize computing costs for NLP. In some cases, a subnetwork picked for one task could be repurposed for another, though Frankle notes this transferability wasn't universal. Still, Frankle is more than happy with the group's results.

"I was kind of shocked this even worked," he says. "It's not something that I took for granted. I was expecting a much messier result than we got."

This discovery of a winning ticket in a BERT model is "convincing," according to Ari Morcos, a scientist at Facebook AI Research. "These models are becoming increasingly widespread," says Morcos. "So it's important to understand whether the lottery ticket hypothesis holds." He adds that the finding could allow BERT-like models to run using far less computing power, "which could be very impactful given that these extremely large models are currently very costly to run."

Frankle agrees. He hopes this work can make BERT more accessible, because it bucks the trend of ever-growing NLP models. "I don't know how much bigger we can go using these supercomputer-style computations," he says. "We're going to have to reduce the barrier to entry." Identifying a lean, lottery-winning subnetwork does just that -- allowing developers who lack the computing muscle of Google or Facebook to still perform cutting-edge NLP. "The hope is that this will lower the cost, that this will make it more accessible to everyone ... to the little guys who just have a laptop," says Frankle. "To me that's really exciting."

make a difference: sponsored opportunity

Story Source:

Materials provided by Massachusetts Institute of Technology. Original written by Daniel Ackerman. Note: Content may be edited for style and length.


Cite This Page:

  • MLA
  • APA
  • Chicago
Massachusetts Institute of Technology. "Shrinking massive neural networks used to model language." ScienceDaily. ScienceDaily, 1 December 2020. <www.sciencedaily.com/releases/2020/12/201201144041.htm>.
Massachusetts Institute of Technology. (2020, December 1). Shrinking massive neural networks used to model language. ScienceDaily. Retrieved December 1, 2020 from www.sciencedaily.com/releases/2020/12/201201144041.htm
Massachusetts Institute of Technology. "Shrinking massive neural networks used to model language." ScienceDaily. www.sciencedaily.com/releases/2020/12/201201144041.htm (accessed December 1, 2020).

  • RELATED TOPICS
    • Mind & Brain
      • Brain-Computer Interfaces
      • Intelligence
      • Language Acquisition
      • Child Development
    • Computers & Math
      • Computer Modeling
      • Computers and Internet
      • Mathematical Modeling
      • Distributed Computing
advertisement

  • RELATED TERMS
    • Computational neuroscience
    • Neural network
    • Computing power everywhere
    • Learning disability
    • Early childhood education
    • Artificial intelligence
    • Mathematical model
    • Information and communication technologies

1

2

3

4

5
RELATED STORIES

'Multitasking' AI Tool to Extract Cancer Data in Record Time
Feb. 12, 2020 — To better leverage cancer data for research, scientists are developing an artificial intelligence (AI)-based natural language processing tool to improve information extraction from textual pathology ...
Accelerating Design, Training of Deep Learning Networks
Jan. 10, 2018 — For deep learning to be effective, existing neural networks to be modified, or novel networks designed and then "trained" so that they know precisely what to look for and can produce valid results. ...
Study Clarifies How Neural Nets 'Think' When Processing Language
Sep. 11, 2017 — A new general-purpose technique has been developed for making sense of neural networks that are trained to perform natural-language-processing tasks, in which computers attempt to interpret freeform ...
Brain-Inspired Device to Power Artificial Systems
Oct. 7, 2016 — New research has demonstrated that a nanoscale device, called a memristor, could be used to power artificial systems that can mimic the human brain. Artificial neural networks (ANNs) exhibit learning ...
FROM AROUND THE WEB

ScienceDaily shares links with sites in the TrendMD network and earns revenue from third-party advertisers, where indicated.
  Print   Email   Share

advertisement

1

2

3

4

5
Most Popular
this week

SPACE & TIME
(c) (c) IM_VISUALS / AdobeField Geology at Mars' Equator Points to Ancient Megaflood
(c) (c) CrispyMedia / AdobeTree Rings May Hold Clues to Impacts of Distant Supernovas on Earth
(c) NASA/JoelNASA's SpaceX Crew-1 Astronauts Headed to International Space Station
MATTER & ENERGY
(c) (c) tampatra / AdobeA Biochemical Random Number
(c) (c) tussik / AdobeTwo Distinctly Different Liquid States of Water
Supersized Wind Turbines Generate Clean Energy--and Surprising Physics
COMPUTERS & MATH
Three Reasons Why COVID-19 Can Cause Silent Hypoxia
Video Games Can Change Your Brain
Researchers Identify Features That Could Make Someone a Virus Super-Spreader
advertisement

Strange & Offbeat
 

SPACE & TIME
Next Step in Simulating the Universe
Experiments Unravelling the Mystery of Mars' Moon Phobos
Puzzling 'Cold Quasar' Forming New Stars in Spite of Active Galactic Nucleus
MATTER & ENERGY
Ultrathin Spray-Applied MXene Antennas Are Ready for 5G
Getting to the Core of Nuclear Speckles
Game Changer in Thermoelectric Materials Could Unlock Body-Heat Powered Personal Devices
COMPUTERS & MATH
More Skin-Like, Electronic Skin That Can Feel
AI System Discovers Useful New Material
World's Smallest Atom-Memory Unit Created
SD
  • SD
    • Home Page
    • Top Science News
    • Latest News
  • Home
    • Home Page
    • Top Science News
    • Latest News
  • Health
    • View all the latest top news in the health sciences,
      or browse the topics below:
      Health & Medicine
      • Allergy
      • Alternative Medicine
      • Birth Control
      • Cancer
      • Diabetes
      • Diseases
      • Heart Disease
      • HIV and AIDS
      • Obesity
      • Stem Cells
      • ... more topics
      Mind & Brain
      • ADD and ADHD
      • Addiction
      • Alzheimer's
      • Autism
      • Depression
      • Headaches
      • Intelligence
      • Psychology
      • Relationships
      • Schizophrenia
      • ... more topics
      Living Well
      • Parenting
      • Pregnancy
      • Sexual Health
      • Skin Care
      • Men's Health
      • Women's Health
      • Nutrition
      • Diet and Weight Loss
      • Fitness
      • Healthy Aging
      • ... more topics
  • Tech
    • View all the latest top news in the physical sciences & technology,
      or browse the topics below:
      Matter & Energy
      • Aviation
      • Chemistry
      • Electronics
      • Fossil Fuels
      • Nanotechnology
      • Physics
      • Quantum Physics
      • Solar Energy
      • Technology
      • Wind Energy
      • ... more topics
      Space & Time
      • Astronomy
      • Black Holes
      • Dark Matter
      • Extrasolar Planets
      • Mars
      • Moon
      • Solar System
      • Space Telescopes
      • Stars
      • Sun
      • ... more topics
      Computers & Math
      • Artificial Intelligence
      • Communications
      • Computer Science
      • Hacking
      • Mathematics
      • Quantum Computers
      • Robotics
      • Software
      • Video Games
      • Virtual Reality
      • ... more topics
  • Enviro
    • View all the latest top news in the environmental sciences,
      or browse the topics below:
      Plants & Animals
      • Agriculture and Food
      • Animals
      • Biology
      • Biotechnology
      • Endangered Animals
      • Extinction
      • Genetically Modified
      • Microbes and More
      • New Species
      • Zoology
      • ... more topics
      Earth & Climate
      • Climate
      • Earthquakes
      • Environment
      • Geography
      • Geology
      • Global Warming
      • Hurricanes
      • Ozone Holes
      • Pollution
      • Weather
      • ... more topics
      Fossils & Ruins
      • Ancient Civilizations
      • Anthropology
      • Archaeology
      • Dinosaurs
      • Early Humans
      • Early Mammals
      • Evolution
      • Lost Treasures
      • Origin of Life
      • Paleontology
      • ... more topics
  • Society
    • View all the latest top news in the social sciences & education,
      or browse the topics below:
      Science & Society
      • Arts & Culture
      • Consumerism
      • Economics
      • Political Science
      • Privacy Issues
      • Public Health
      • Racial Disparity
      • Religion
      • Sports
      • World Development
      • ... more topics
      Business & Industry
      • Biotechnology & Bioengineering
      • Computers & Internet
      • Energy & Resources
      • Engineering
      • Medical Technology
      • Pharmaceuticals
      • Transportation
      • ... more topics
      Education & Learning
      • Animal Learning & Intelligence
      • Creativity
      • Educational Psychology
      • Educational Technology
      • Infant & Preschool Learning
      • Learning Disorders
      • STEM Education
      • ... more topics
  • Quirky
    • Top News
    • Human Quirks
    • Odd Creatures
    • Bizarre Things
    • Weird World
Free Subscriptions

Get the latest science news with ScienceDaily's free email newsletters, updated daily and weekly. Or view hourly updated newsfeeds in your RSS reader:

  • Email Newsletters
  • RSS Feeds
Follow Us

Keep up to date with the latest news from ScienceDaily via social networks:

  • Facebook
  • Twitter
  • LinkedIn
Have Feedback?

Tell us what you think of ScienceDaily -- we welcome both positive and negative comments. Have any problems using the site? Questions?

  • Leave Feedback
  • Contact Us
About This Site  |  Staff  |  Reviews  |  Contribute  |  Advertise  |  Privacy Policy  |  Editorial Policy  |  Terms of Use
Copyright 2020 ScienceDaily or by other parties, where indicated. All rights controlled by their respective owners.
Content on this website is for information only. It is not intended to provide medical or other professional advice.
Views expressed here do not necessarily reflect those of ScienceDaily, its staff, its contributors, or its partners.
Financial support for ScienceDaily comes from advertisements and referral programs, where indicated.
— CCPA: Do Not Sell My Information — — GDPR: Privacy Settings —