DeepMind Surges on, Releasing Acme and Reverb RL Libraries

栏目: IT技术 · 发布时间: 5年前

内容简介:The Alphabet subsidiary continues to pump out useful software libraries for the machine learning research community.Deepmind, now wholly an Alphabeta subsidiary, is an innovating software company focusing on Artificial Intelligence. You likely know of them

DeepMind Surges on, Releasing Acme and Reverb RL Libraries

The Alphabet subsidiary continues to pump out useful software libraries for the machine learning research community.

The Gist

Deepmind, now wholly an Alphabeta subsidiary, is an innovating software company focusing on Artificial Intelligence. You likely know of them due to their accomplishments in training AlphaGo and then AlphaGoZero . The AlphaGoZero reinforcement learning agent learned largely from scratch to become the Go world champion .

Despite this and other radical successes, the company’s academic achievements have been faced with frustrations, due to a lack of reproducibility. The ability to independently reproduce academic works is the lifeblood for validation and further collaboration. Thus, it is crucial to provide, especially in a world of increasing divide between the compute-rich and everyone else.

DeepMind understands. They have gotten into a groove in recent years, consistently releasing modular software libraries to aid fellow researchers. These libraries have served numerous purposes, including the following:

ReproducibilitySimplicityModularityParallelizationEfficiency

With the release of their Acme and Reverb libraries, this trend continues nicely. In fact, the authors of the library explicitly call out the high-level goals of Acme on their website:

1. To enable the reproducibility of our methods and results — this will help clarify what makes an RL problem hard or easy, something that is seldom apparent.

2. To simplify the way we (and the community at large) design new algorithms — we want that next RL agent to be easier for everyone to write!

3. To enhance the readability of RL agents — there should be no hidden surprises when transitioning from a paper to code.

DeepMind Acme Authors

Ok, But How?

One of the ways in which they achieve these goals is through appropriate levels of abstraction. The field of reinforcement learning is like an onion, in that its best used in terms of its layers. At face value, you have an agent that learns from data. Peeling back the data part, you see that this data is either a stored data set or a live sequence of experiences. Peeling back the agent, you see that it plans and takes actions, resulting in a measured response from its environment. You can again peel back more, and delve into policies, experience, replay, etc. The illustration below shows this nicely.

A hierarchical display of the reinforcement learning problem

Another way in which Acme achieves its goals is via a scalable data storage mechanism, implemented as the companion Reverb library. To motivate this, consider the typical experience replay buffer for an agent. How big does that buffer get? Usually, it’s on the order of at least tens to hundreds of thousands of experience tuples, and that’s per-agent. When working on a simulation involving thousands to millions of agents, you get….a lot.

By de-coupling, the notions of data producers (agents) and data consumers (learners), an efficient data storage mechanism can sit independently between the two. This is exactly what Reverb accomplishes. As a company-backed library with over 70% code in C++ and a neat python interface on top, I’m really excited to dive deep in this one.

https://www.youtube.com/watch?v=3hnlDfJYWcI&feature=youtu.be

An R2D2 RL agent playing the arcade game Breakout

Conclusion

By continuously releasing fantastic open source libraries like these, DeepMind helps to lower the barrier for entry and level the playing field for research in ML and AI. Pair this with low-cost cloud computing solutions, and anyone can jump right in! Send in any cool projects you make with these libraries. I can’t wait to see them.

Jump In

Stay Up To Date

Things move quickly in academics and industry! Keep yourself updated with the general LifeWithData blog as well as the ML UTD newsletter.

If you’re not a fan of newsletters, but still want to stay in the loop, consider adding lifewithdata.org/blog and lifewithdata.org/tag/ml-utd to a Feedly aggregation setup.


以上就是本文的全部内容,希望对大家的学习有所帮助,也希望大家多多支持 码农网

查看所有标签

猜你喜欢:

本站部分资源来源于网络,本站转载出于传递更多信息之目的,版权归原作者或者来源机构所有,如转载稿涉及版权问题,请联系我们

风云Flash ActionScript高级编程艺术

风云Flash ActionScript高级编程艺术

赵英杰 / 第1版 (2006年7月1日) / 2006-7 / 45.00元

本书从基本的Actionscript语言概念开始介绍,配以实际的程序实例并穿插生动的图示说明,深入浅出地讲解Flash ActionScript程序的运用逻辑与概念,让读者从实例中学习进而融会贯通。同时,本书也说明面向对象程序设计(00P)的语法及常用类别实例,提升读者制作F1ash作品的造诣和能力。全书共分为10章,精彩实例包括以三角函数制作的抽奖轮盘,FlashLite手机版孔明棋游戏,Bit......一起来看看 《风云Flash ActionScript高级编程艺术》 这本书的介绍吧!

MD5 加密
MD5 加密

MD5 加密工具

SHA 加密
SHA 加密

SHA 加密工具

XML 在线格式化
XML 在线格式化

在线 XML 格式化压缩工具