• Home
  • About Us
  • Contact Us
  • Privacy Policy
  • Special Offers
Business Intelligence Info
  • Business Intelligence
    • BI News and Info
    • Big Data
    • Mobile and Cloud
    • Self-Service BI
  • CRM
    • CRM News and Info
    • InfusionSoft
    • Microsoft Dynamics CRM
    • NetSuite
    • OnContact
    • Salesforce
    • Workbooks
  • Data Mining
    • Pentaho
    • Sisense
    • Tableau
    • TIBCO Spotfire
  • Data Warehousing
    • DWH News and Info
    • IBM DB2
    • Microsoft SQL Server
    • Oracle
    • Teradata
  • Predictive Analytics
    • FICO
    • KNIME
    • Mathematica
    • Matlab
    • Minitab
    • RapidMiner
    • Revolution
    • SAP
    • SAS/SPSS
  • Humor

UC Berkeley researchers open-source RAD to improve any reinforcement learning algorithm

May 2, 2020   Big Data
 UC Berkeley researchers open source RAD to improve any reinforcement learning algorithm

A group of University of California, Berkeley researchers this week open-sourced Reinforcement Learning with Augmented Data (RAD). In an accompanying paper, the authors say this module can improve any existing reinforcement learning algorithm and that RAD achieves better compute and data efficiency than Google AI’s PlaNet, as well as recently released cutting-edge algorithms like DeepMind’s Dreamer and SLAC from UC Berkeley and DeepMind.

RAD achieves state-of-the-art results on common benchmarks and matches or beats every baseline in terms of performance and data efficiency across 15 DeepMind control environments, the researchers say. It does this in part by applying data augmentations for visual observations. Coauthors of the paper on RAD include Michael “Misha” Laskin, Kimin Lee, and Berkeley AI Research codirector and Covariant founder Pieter Abbeel.

RAD was released Thursday on preprint repository arXiv. Data augmentation has been important to advances in convolutional neural networks (CNN) for challenges like robotic grasping and achieving human-level performance in games like Go.

“For the first time, we show that data augmentations alone can significantly improve the data-efficiency and generalization of RL methods operating from pixels, without any changes to the underlying RL algorithm, on the DeepMind Control Suite and the OpenAI ProcGen benchmarks, respectively,” the paper reads. “By using multiple augmented views of the same data point as input, CNNs are forced to learn consistencies in their internal representations. This results in a visual representation that improves generalization, data-efficiency, and transfer learning.”

VB Transform 2020 Online – July 15-17, 2020: Join leading AI executives at VentureBeat’s AI event of the year. Register today and save 30% off digital access passes.

Data augmentation techniques increase diversity in training data sets without collecting new data. “We find that data diversity alone can make agents focus on meaningful information from high-dimensional observations without any changes to the reinforcement learning method,” the authors note.

It’s been a busy week for the machine learning subfield of reinforcement learning.

Earlier this week, NYU researchers released work on arXiv that applies data augmentation they say also achieves state-of-the-art results on the DeepMind control suite.

And at the entirely digital International Conference on Learning Representations (ICLR) this week, Google AI researchers introduced methods for measuring the reliability of reinforcement learning algorithms, and Huawei AI researchers introduced Adversarial AutoAugment for improving data augmentation policy.

Abbeel also coauthored a number of reinforcement algorithm papers at ICLR, including HiPPO for training several levels of reinforcement learning algorithms at once and a paper on reinforcement learning and policy optimization that touches on data augmentation.

In a different series of developments, earlier this week Salesforce released the AI Economist, reinforcement learning the company claims is able to create optimal tax policies.

Let’s block ads! (Why?)

Big Data – VentureBeat

Algorithm, Berkeley, Improve, Learning, opensource, reinforcement, researchers
  • Recent Posts

    • Kevin Hart Joins John Hamburg For New Netflix Comedy Film Titled ‘Me Time’
    • Who is Monitoring your Microsoft Dynamics 365 Apps?
    • how to draw a circle using disks, the radii of the disks are 1, while the radius of the circle is √2 + √6
    • Tips on using Advanced Find in Microsoft Dynamics 365
    • You don’t tell me where to sit.
  • Categories

  • Archives

    • February 2021
    • January 2021
    • December 2020
    • November 2020
    • October 2020
    • September 2020
    • August 2020
    • July 2020
    • June 2020
    • May 2020
    • April 2020
    • March 2020
    • February 2020
    • January 2020
    • December 2019
    • November 2019
    • October 2019
    • September 2019
    • August 2019
    • July 2019
    • June 2019
    • May 2019
    • April 2019
    • March 2019
    • February 2019
    • January 2019
    • December 2018
    • November 2018
    • October 2018
    • September 2018
    • August 2018
    • July 2018
    • June 2018
    • May 2018
    • April 2018
    • March 2018
    • February 2018
    • January 2018
    • December 2017
    • November 2017
    • October 2017
    • September 2017
    • August 2017
    • July 2017
    • June 2017
    • May 2017
    • April 2017
    • March 2017
    • February 2017
    • January 2017
    • December 2016
    • November 2016
    • October 2016
    • September 2016
    • August 2016
    • July 2016
    • June 2016
    • May 2016
    • April 2016
    • March 2016
    • February 2016
    • January 2016
    • December 2015
    • November 2015
    • October 2015
    • September 2015
    • August 2015
    • July 2015
    • June 2015
    • May 2015
    • April 2015
    • March 2015
    • February 2015
    • January 2015
    • December 2014
    • November 2014
© 2021 Business Intelligence Info
Power BI Training | G Com Solutions Limited