查看原文
其他

ICLR 2019热议论文Top 5:BigGAN、斗地主深度学习算法等

专知 新智元 2018-12-28

新智元 AI World 2018 世界人工智能峰会

全程回顾

新智元于9月20日在北京国家会议中心举办AI WORLD 2018世界人工智能峰会,邀请机器学习教父、CMU教授 Tom Mitchell,迈克思·泰格马克,周志华,陶大程,陈怡然等AI领袖一起关注机器智能与人类命运。


爱奇艺

上午:https://www.iqiyi.com/v_19rr54cusk.html

下午:https://www.iqiyi.com/v_19rr54hels.html

新浪:http://video.sina.com.cn/l/p/1724373.html







  新智元推荐 

来源:专知(ID: Quan_Zhuanzhi)

【新智元导读】国庆佳节第一天,举国同庆!出门旅游想必到处都是人山人海,不如在家里看看论文也是极好的!近日,机器学习顶会之一的ICLR2019投稿刚刚截止,本次大会投稿论文采用匿名公开的方式。本文整理了目前在国外社交网络和知乎上讨论热烈的一些论文,一起来看看!


首先我们来看看ICLR 2018,也就是去年的提交论文题目分布情况。如下图所示。热门关键词:强化学习、GAN、RIP等。



上图为ICLR 2019提交论文的分布情况,热门关键词:强化学习、GAN、元学习等等。可以看出比去年还是有些变化的。


投稿论文地址:

https://openreview.net/group?id=ICLR.cc/2019/Conference




在Google Colaboratory上可以找到关于ICLR 2019提交论文话题之间更加直观的可视化图。我们选择了上图中排名第三的话题“GAN”,图中由红色表示。可以看出,排名第三的GAN与表中多个话题有交集,如training、state、graph等。


讨论热度最高的论文TOP 5


1. LARGE SCALE GAN TRAINING FOR HIGH FIDELITY NATURAL IMAGE SYNTHESIS

最强GAN图像生成器,真假难辨

论文地址:

https://openreview.net/pdf?id=B1xsqj09Fm


更多样本地址:

https://drive.google.com/drive/folders/1lWC6XEPD0LT5KUnPXeve_kWeY-FxH002


第一篇就是这篇最佳BigGAN,DeepMind负责星际项目的Oriol Vinyals,说这篇论文带来了史上最佳的GAN生成图片,提升Inception Score 100分以上。


论文摘要:


尽管近期由于生成图像建模的研究进展,从复杂数据集例如 ImageNet 中生成高分辨率、多样性的样本仍然是很大的挑战。为此,研究者尝试在最大规模的数据集中训练生成对抗网络,并研究在这种规模的训练下的不稳定性。研究者发现应用垂直正则化(orthogonal regularization)到生成器可以使其服从简单的「截断技巧」(truncation trick),从而允许通过截断隐空间来精调样本保真度和多样性的权衡。这种修改方法可以让模型在类条件的图像合成中达到当前最佳性能。当在 128x128 分辨率的 ImageNet 上训练时,本文提出的模型—BigGAN—可以达到 166.3 的 Inception 分数(IS),以及 9.6 的 Frechet Inception 距离(FID),而之前的最佳 IS 和 FID 仅为 52.52 和 18.65。

BigGAN的生成器架构

生成样例,真是惟妙惟肖


2. Recurrent Experience Replay in Distributed Reinforcement Learning

分布式强化学习中的循环经验池

论文地址:

https://openreview.net/pdf?id=r1lyTjAqYX


Building on the recent successes of distributed training of RL agents, in this paper we investigate the training of RNN-based RL agents from experience replay. We investigate the effects of parameter lag resulting in representational drift and recurrent state staleness and empirically derive an improved training strategy. Using a single network architecture and fixed set of hyper-parameters, the resulting agent, Recurrent Replay Distributed DQN, triples the previous state of the art on Atari-57, and surpasses the state of the art on DMLab-30. R2D2 is the first agent to exceed human-level performance in 52 of the 57 Atari games.



3. Shallow Learning For Deep Networks

深度神经网络的浅层学习

论文地址:

https://openreview.net/forum?id=r1Gsk3R9Fm



浅层监督的一层隐藏层神经网络具有许多有利的特性,使它们比深层对应物更容易解释,分析和优化,但缺乏表示能力。在这里,我们使用1-hiddenlayer学习问题逐层顺序构建深层网络,这可以从浅层网络继承属性。与之前使用浅网络的方法相反,我们关注的是深度学习被认为对成功至关重要的问题。因此,我们研究了两个大规模图像识别任务的CNN:ImageNet和CIFAR-10。使用一组简单的架构和训练想法,我们发现解决序列1隐藏层辅助问题导致CNN超过ImageNet上的AlexNet性能。通过解决2层和3层隐藏层辅助问题来扩展ourtraining方法以构建单个层,我们获得了一个11层网络,超过ImageNet上的VGG-11,获得了89.8%的前5个单一作物。据我们所知,这是CNN的端到端培训的第一个竞争性替代方案,可以扩展到ImageNet。我们进行了广泛的实验来研究它在中间层上引起的性质。



4. Relational Graph Attention Networks

关联性图注意力网络

论文地址:

https://openreview.net/forum?id=Bklzkh0qFm&noteId=HJxMHja3Y7

论文摘要:

In this paper we present Relational Graph Attention Networks, an extension of Graph Attention Networks to incorporate both node features and relational information into a masked attention mechanism, extending graph-based attention methods to a wider variety of problems, specifically, predicting the properties of molecules. We demonstrate that our attention mechanism gives competitive results on a molecular toxicity classification task (Tox21), enhancing the performance of its spectral-based convolutional equivalent. We also investigate the model on a series of transductive knowledge base completion tasks, where its performance is noticeably weaker. We provide insights as to why this may be, and suggest when it is appropriate to incorporate an attention layer into a graph architecture.


5. A Solution to China Competitive Poker Using Deep Learning

斗地主深度学习算法

论文地址:

https://openreview.net/forum?id=rJzoujRct7


论文摘要:

Recently, deep neural networks have achieved superhuman performance in various games such as Go, chess and Shogi. Compared to Go, China Competitive Poker, also known as Dou dizhu, is a type of imperfect information game, including hidden information, randomness, multi-agent cooperation and competition. It has become widespread and is now a national game in China. We introduce an approach to play China Competitive Poker using Convolutional Neural Network (CNN) to predict actions. This network is trained by supervised learning from human game records. Without any search, the network already beats the best AI program by a large margin, and also beats the best human amateur players in duplicate mode.


其他有意思的论文:


ICLR 2019 有什么值得关注的亮点?- 周博磊的回答 - 知乎 

https://www.zhihu.com/question/296404213/answer/500575759


问句开头式:


  • Are adversarial examples inevitable?

  • Transfer Value or Policy? A Value-centric Framework Towards Transferrable Continuous Reinforcement Learning

  • How Important is a Neuron?

  • How Powerful are Graph Neural Networks?

  • Do Language Models Have Common Sense?

  • Is Wasserstein all you need?


哲理警句式:


  • Learning From the Experience of Others: Approximate Empirical Bayes in Neural Networks

  • In Your Pace: Learning the Right Example at the Right Time

  • Learning what you can do before doing anything

  • Like What You Like: Knowledge Distill via Neuron Selectivity Transfer

  • Don’s Settle for Average, Go for the Max: Fuzzy Sets and Max-Pooled Word Vectors


抖机灵式:


  • Look Ma, No GANs! Image Transformation with ModifAE

  • No Pressure! Addressing Problem of Local Minima in Manifold Learning

  • Backplay: 'Man muss immer umkehren'

  • Talk The Walk: Navigating Grids in New York City through Grounded Dialogue

  • Fatty and Skinny: A Joint Training Method of Watermark

  • A bird's eye view on coherence, and a worm's eye view on cohesion

  • Beyond Winning and Losing: Modeling Human Motivations and Behaviors with Vector-valued Inverse Reinforcement Learning


一句总结式:


  • ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness.


本文经授权转载自微信公众号“专知”(ID: Quan_Zhuanzhi)


新智元AI WORLD 2018

世界人工智能峰会全程回顾


新智元于9月20日在北京国家会议中心举办AI WORLD 2018世界人工智能峰会,邀请机器学习教父、CMU教授 Tom Mitchell,迈克思·泰格马克,周志华,陶大程,陈怡然等AI领袖一起关注机器智能与人类命运。


全程回顾新智元 AI World 2018 世界人工智能峰会盛况: 

爱奇艺 

上午:https://www.iqiyi.com/v_19rr54cusk.html 

下午:https://www.iqiyi.com/v_19rr54hels.html

新浪:http://video.sina.com.cn/l/p/1724373.html

    您可能也对以下帖子感兴趣

    文章有问题?点此查看未经处理的缓存