(转) Reinforcement Learning for Profit
Reinforcement Learning for Profit
July 17, 2016
Is RL being used in revenue generating systems today?
Recently, one of my facebook friends, and alumni of the University of Alberta (with a PhD in Computing Science), Cosmin Paduraru posed a question:
Where is Reinforcement Learning used in revenue generating systems today?
I have been thinking about this lots over the last month as I attended two international conferences on Artificial Intelligence and Machine Learning (ICML and IJCAI) in NYC, USA. It is important to explore future prospects both inside and outside academia — In case you need a catch up, I am currently at the University of Alberta working on a PhD in Computing Science with a focus on Reinforcement Learning and Artificial Intelligence.
With the success of modern AI systems — out of the winter and into the spring — many companies have invested and continue to invested heavily into modern AI systems, backed by teams of leading researchers in the field (e.g. Facebook, Google, Microsoft, IBM, Twitter, etc.).
With that said, maybe Cosmin is right, Reinforcement Learning (Sutton and Barto 1998, and this killer-intro by the fantastically talented Andrej Karpathy) is seemingly publicly underrepresented in currently deployed systems making money in the real world, or is it?
Adapted from Sutton and Barto 1998 and WALL-E
Luckily I was at the International Joint Conference on Artificial Intelligence where I was attending a panel discussion on The Business of AI, the panel was composed of all speakers from the industry day. A desirable venue to solicit a wide variety of opinions from thought leaders in the field.
So I posed the question to them, their responses went as follows:
Peter Norvig (Director of Research at Google): “well… AlphaGo made a million bucks and then gave it away”… a recent tweet from Demis Hassabis (Google DeepMind) confirms:
Pleased to confirm the recipients of the #AlphaGo $1m prize! @UNICEF_uk, @CodeClub, and the American, European and Korean Go associations
— Demis Hassabis (@demishassabis) June 6, 2016
Peter Stone (Founder and President, Cogitai. Professor UoT (Austin)) gave lots of great examples of recent applications:
He said,“We are on the cusp of moving from the academic lab to the industry for RL, adaptation, and lifelong learning…We are at the cusp, and that is the main motivation from Cogitai”
He also referenced work by Thomas G. Dietterich on invasive species management, wildfire suppression, by Joelle Pineau on applying RL in healthcare, and by Andrew Ng and Drew Bagnall on helicopter control. All of these could be as a practical demonstrations of specific, developing industrial applications.
Hiroaki Kitano (President & CEO SONY Computer Science Laboratories) said that this is a current research area for Sony and to expect profitability using these and advancing RL algorithms in 2-5 years. Almost 10 years after Sony’s last robotic venture, the Aibo, Sony CEO Kazuo Hirai has just recently (late June 2016) said “the robots we are developing can have emotional bonds with customers, giving them joy and becoming the objects of love”.
Guruduth Banavar (Chief Science Officer, Cognitive Computing, IBM Research) predicted that this is going to happen, sooner rather than later, and his prediction was that it will happen in the domain of conversational systems, dialog systems, and understanding the larger context of conversations. He also mentioned that the illustrious Gerald Tesauro (the man behind TD-Gammon) is working on these problems. Interesting that he did not mention Watson…
Some interesting answers from industry leaders. But I was surprised that no one mentioned:recommender systems (like those on Amazon, Netflix, Yelp, and nicely formalized as an RL problem in 2005 by Shani et al.), are these systems all collaborative filtering? Surely not.
No one mentioned that Google Reinforcement Learning Architecture (here is a quick summary), which I can only imagine could be behind some of the personal recommendations and rankings that Google does behind-the-scenes on Search, YouTube, and maybe … Maps?
No one mentioned contextual bandits, sometimes called associative RL (as discussed by Li et al. 2010 for news recommendation), for serving ads and news stories. These systems are surely deployed on large-scale news sites by the publishers to maximize click-through-ratios and create a personalized experience. Microsoft recently announced Multiworld Testing Decision Service, for making context based decisions… I guess there were no Microsoft representatives on the panel to toot this horn (thanks for the catch Pardis)
With so much potentially out there, why was there no mention of these use cases for reinforcement learning? Where else could RL be hiding in the money-making wild? RL seems like an ideal candidate for systems of personalization on large-scale, sequential decision-making problems… so what am I missing?
(转) Reinforcement Learning for Profit的更多相关文章
- [转]Introduction to Learning to Trade with Reinforcement Learning
Introduction to Learning to Trade with Reinforcement Learning http://www.wildml.com/2018/02/introduc ...
- Introduction to Learning to Trade with Reinforcement Learning
http://www.wildml.com/2015/12/implementing-a-cnn-for-text-classification-in-tensorflow/ The academic ...
- Machine Learning Algorithms Study Notes(5)—Reinforcement Learning
Reinforcement Learning 对于控制决策问题的解决思路:设计一个回报函数(reward function),如果learning agent(如上面的四足机器人.象棋AI程序)在决定 ...
- (转) Playing FPS games with deep reinforcement learning
Playing FPS games with deep reinforcement learning 博文转自:https://blog.acolyer.org/2016/11/23/playing- ...
- (zhuan) Deep Reinforcement Learning Papers
Deep Reinforcement Learning Papers A list of recent papers regarding deep reinforcement learning. Th ...
- (转) Deep Learning Research Review Week 2: Reinforcement Learning
Deep Learning Research Review Week 2: Reinforcement Learning 转载自: https://adeshpande3.github.io/ad ...
- Learning Roadmap of Deep Reinforcement Learning
1. 知乎上关于DQN入门的系列文章 1.1 DQN 从入门到放弃 DQN 从入门到放弃1 DQN与增强学习 DQN 从入门到放弃2 增强学习与MDP DQN 从入门到放弃3 价值函数与Bellman ...
- Open source packages on Deep Reinforcement Learning
智能车 self driving car + 强化学习 reinforcement learning + 神经网络 模拟 https://github.com/MorvanZhou/my_resear ...
- (转) Deep Reinforcement Learning: Playing a Racing Game
Byte Tank Posts Archive Deep Reinforcement Learning: Playing a Racing Game OCT 6TH, 2016 Agent playi ...
随机推荐
- 一道面试题,简单模拟spring ioc
自己实现的,程序写的土了点,很多情况没去考虑,主要是复习理解怎么使用反射来实现spring 的依赖注入. package dom4jtest; import java.lang.reflect.Inv ...
- return, exit, _exit的区别
return是返回的最常用的方式 _exit属于POSIX定义的系统调用 exit是GLIBC封装之后的函数 1 _exit和exit都会导致整个进程退出,清理进程所占用的资源,但是glibc封装ex ...
- Design Patterns---- Strategy 模式
设计模式:可复用面向对象软件的基础 书中对 Strategy 模式的定义如下: 定义了一系列的算法,把它们一个个封装起来,并且使它们可相互替换.本模式使得算法可独立于它的用户而变化. 案例:设计一个商 ...
- PHP图像处理之在原图像处理
处理原有的图像 图片处理,缩放,裁剪,翻转,旋转,透明,锐化等图片操作 一.创建图片资源 imagecreatetruecolor(width,hei ...
- PHP中的数组(二)常用数组处理函数
数组的相关处理函数 一.数组键/值操作有关的函数 1.array_values() 无论是关联的还是索引的返回的都是索引数组 <?php $lamp=array(&quo ...
- PHP_Const
PHP_Const 常量 规则: 1 总是大写 2 A-Z 及 从127-255的ASCII字符 3 全局范围 4 用define函数定义 5 只能包含标量数据 如Boolean integer fl ...
- Sheet can not be presented because the view is not in a window的解决办法,和window的简单使用
Sheet can not be presented because the view is not in a window,顺便在stackoverflow上找了答案,希望能给大家带来帮助,在此感谢 ...
- 10、C#基础整理(集合)
集合 1.集合的引用 using System.Collections;//添加类 2.定义集合(ArrayList 或 Array) ArrayList arr = new ArrayList(); ...
- HDU 5823 (状压dp)
Problem color II 题目大意 定义一个无向图的价值为给每个节点染色使得每条边连接的两个节点颜色不同的最少颜色数. 对于给定的一张由n个点组成的无向图,求该图的2^n-1张非空子图的价值. ...
- LCA(倍增)
type arr=record v,nt:longint; end; ; lx=; ..maxn] of longint; eg:..maxn*] of arr; d:..maxn] of longi ...