[robotics-worldwide] [meetings] CfP: Goal Specifications for Reinforcement Learning @ FAIM 2018

classic Classic list List threaded Threaded
1 message Options
Reply | Threaded
Open this post in threaded view
|

[robotics-worldwide] [meetings] CfP: Goal Specifications for Reinforcement Learning @ FAIM 2018

aedwards8
Hi,

We would like to invite submissions for the 1st Workshop on Goal
Specifications for Reinforcement Learning at the Federated AI Meeting 2018.
The submission deadline is May 1st 2018. Attached below is the call for
papers. We look forward to your submissions.

======================================================================
ICML/IJCAI/AAMAS 2018 Workshop: Goal Specifications for Reinforcement
Learning
Stockholm, Sweden
https://urldefense.proofpoint.com/v2/url?u=https-3A__sites.google.com_view_goalsrl&d=DwICAg&c=clK7kQUTWtAVEOVIgvi0NU5BOUHhpN0H8p7CSfnc_gI&r=0w3solp5fswiyWF2RL6rSs8MCeFamFEPafDTOhgTfYI&m=4P8QGiPEAwnru3X8ngDLwUIf3NcD932X6mgTJZ8KEUk&s=B9FlA8EGA9ITQxecS3JB_qQXwkKCg7oj6uMRJcqwZOQ&e=
======================================================================

==================
IMPORTANT DATES
==================
Paper submission opens: April 1st, 2018
Submission deadline: May 1st, 2018
Author notification: June 1st, 2018
Camera-ready deadline: June 21st, 2018
Workshop: July 13/14/15th, 2018

==================
ABSTRACT
==================
Reinforcement Learning (RL) agents traditionally rely on hand-designed
scalar rewards to learn how to act. The more complex and diverse
environments and tasks become, the more difficult it may be to engineer
rewards that elicit desired behavior. Designing rewards in multi-agent
settings with adversaries or co-operative allies can be even more
complicated. Experiment designers often have a goal in mind and then must
reverse engineer a reward function that will likely lead to it. This process
can be difficult, especially for non-experts, and is susceptible to reward
hacking---unexpected and undesired behavior that achieves high reward but
does not capture the essence of what the engineer was trying to achieve.
Moreover, hand-designed reward functions may be brittle, as slight changes
in the environment may yield large, and potentially unsafe, alterations in
agent behavior.

The community has addressed these problems through many disparate approaches
including reward shaping, intrinsic rewards, hierarchical reinforcement
learning, curriculum learning, and transfer learning. Another approach is to
avoid designing scalar rewards altogether, and rather focus on designing
goals, for example, through inverse reinforcement learning, imitation
learning, target images, or multimodal channels such as speech and text.

This workshop will consider all topics related to designing goals for
reinforcement learning and problems that can arise from ill-defined goals.
The submissions can include novel research, open problems in the field, and
surveys. We are particularly interested in the topics of reward engineering,
reward hacking, interpretability, learning from humans and goal design using
multimodal input.


==================
AREAS OF INTEREST
==================
Problems with reward design
        - Robust reward functions
        - Reward hacking
        - Adversarial attacks on RL agents
        - Generalizability of reward functions
        - Communicating learned goals to humans

Methods of reward design
        - Reward engineering
        - Reward shaping
        - Intrinsic rewards

Methods of learning rewards
        - Inverse Reinforcement Learning
        - Interactive learning
        - Supervised learning
        - Evolutionary approaches

Methods of goal design using:
        - Target images
        - Imitation learning
        - Transfer learning
        - Curriculum learning
        - Hierarchical RL
        - Multimodal input (speech, text, sketches, etc.)
        - Multi-agent cooperative/competitive learning
        - Application-related issues and solutions

==================
SUBMISSION
==================
Submissions will be double-blind and are limited to 4 pages for short papers
and 8 pages for full papers, not including references and appendices.
Formatting should be in ICML style. Concurrent submissions are allowed, but
works that have been accepted at archival venues are discouraged.

Submission link: https://urldefense.proofpoint.com/v2/url?u=https-3A__easychair.org_conferences_-3Fconf-3Dgoalsrl2018&d=DwICAg&c=clK7kQUTWtAVEOVIgvi0NU5BOUHhpN0H8p7CSfnc_gI&r=0w3solp5fswiyWF2RL6rSs8MCeFamFEPafDTOhgTfYI&m=4P8QGiPEAwnru3X8ngDLwUIf3NcD932X6mgTJZ8KEUk&s=Q72_Xulr5Ffn9K-bqZT3BTcJ74j8hbJX9a6lBBKAZMI&e=

==================
ORGANIZERS
==================
Ashley Edwards, Georgia Institute of Technology
Himanshu Sahni, Georgia Institute of Technology
Kaushik Subramanian, Cogitai
Charles Isbell, Georgia Institute of Technology
Michael Littman, Brown University

==================
CONTACT
==================
Please address questions to: [hidden email]




--
Sent from: https://urldefense.proofpoint.com/v2/url?u=http-3A__robotics-2Dworldwide.1046236.n5.nabble.com_&d=DwICAg&c=clK7kQUTWtAVEOVIgvi0NU5BOUHhpN0H8p7CSfnc_gI&r=0w3solp5fswiyWF2RL6rSs8MCeFamFEPafDTOhgTfYI&m=4P8QGiPEAwnru3X8ngDLwUIf3NcD932X6mgTJZ8KEUk&s=Z6eDsdGpgvko672uye6O5q_Up0TMC6JubcGZXqO1RfY&e=
_______________________________________________
robotics-worldwide mailing list
[hidden email]
http://duerer.usc.edu/mailman/listinfo.cgi/robotics-worldwide