Data Driven Network Control with Reinforcement Learning-十大博彩公司
          • <i id="ehhce"><rp id="ehhce"><acronym id="ehhce"></acronym></rp><q id="ehhce"><video id="ehhce"><area id="ehhce"></area></video></q></i><code id="ehhce"></code>
          • <output id="ehhce"></output><noscript id="ehhce"><rt id="ehhce"><output id="ehhce"></output><i id="ehhce"><li id="ehhce"><cite id="ehhce"></cite></li><rp id="ehhce"><dt id="ehhce"></dt></rp></i></rt></noscript>

            <tbody id="ehhce"><input id="ehhce"><object id="ehhce"></object></input></tbody>

          • <aside id="ehhce"><aside id="ehhce"></aside></aside>
          • 文章详情
            您所在位置:首页 > 学术交流 > Data Driven Network Control with Reinforcement Learning

            Data Driven Network Control with Reinforcement Learning

            2018-12-13来源: 浏览次数:

            题目:Data Driven Network Control with Reinforcement Learning



            报告人:崔曙光教授 香港中文大学(深圳)


            We first start with a brief introduction of Reinforcement Learning (RL) and then discuss its applications in self-organizing networks. The first application is on handover control: We propose a two-layer framework to learn the optimal HO controllers in possibly large-scale wireless systems supporting mobile users, where the user mobility patterns could be heterogeneous. In particular, our proposed framework first partitions the User Equipments (UEs) with different mobility patterns into clusters, where the mobility patterns are similar in the same cluster. Then, within each cluster, an asynchronous multi-user deep RL scheme is developed to control the HO processes across the UEs in each cluster, in the goal of lowering the HO rate while ensuring certain system throughput. At each user, a deep-RL framework with LSTM RNN is used. We show that the adopted global-parameter-based asynchronous framework enables us to train faster with more UEs, which could nicely address the scalability issue to support large systems. The second application is on joint energy and access control in energy harvesting wireless systems, where we show that a double-deep-RL solution could lead us to significant system gains.



            崔曙光的研究论文被广泛地引用,在2014年当选为Thomson Reuters高被引科学家,并被Sciencewatch列为世界最具影响力科学家之一。他荣获了IEEE Signal Processing Society 2012最佳论文奖,也是两次最佳会议论文的获得者。他一直在担任多个专业会议、期刊和委员会的主席、分区主编或副主编。他在2013年当选了IEEE Fellow,并在2014年当选IEEE通信协会杰出讲师。2018年入选教育部长江学者和广东省珠江创新团队带头人。

            雷达信号处理国家级重点实验室 信息与通信工程学部

            雷达认知探测成像识别“111基地” 国际合作与交流处

          • 上一篇文章:Analysis of Alternating Direction Method of Multipliers for Nonconvex Problems
          • 下一篇文章:华盛顿大学Jenq-Neng Hwang教授访问足球十大博彩公司开展学术交流
          • 返回顶部

            十大博彩公司 |香港十大博彩公司 |足球十大博彩公司 | |手机版 | | 私彩平台|私彩平台|私彩平台|十大博彩公司|十大博彩公司|十大博彩公司|