LIBRARIES
    • Login
    Research Exchange
    Share your work
    View Item 
    •   Research Exchange
    • Electronic Dissertations and Theses
    • Electronic Dissertations
    • View Item
    •   Research Exchange
    • Electronic Dissertations and Theses
    • Electronic Dissertations
    • View Item
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Browse

    All of Research ExchangeCommunities & CollectionsBy Issue DateAuthorsTitlesSubjectsThis CollectionBy Issue DateAuthorsTitlesSubjects

    My Account

    LoginRegister

    Knowledge Transfer in Reinforcement Learning: How agents should benefit from prior knowledge

    Thumbnail
    View/Open
    Wang_wsu_0251E_12924.pdf (5.939Mb)
    Date
    2019
    Author
    Wang, Zhaodong
    Metadata
    Show full item record
    Abstract
    Reinforcement learning (RL) has had many successes in different tasks, but in practice, it often requires significant amounts of data or training time to learn high-performing policies. For complicated tasks, the learning speed may be too slow to be feasible. To speed up the learning, external prior knowledge could be leveraged by the RL agent. Knowledge transfer could be performed to allow a trained (source) agent to assist a new (target) agent. It is also possible that a human demonstrator could behave as the source. With help from the prior, the new agent is expected to learn faster than learning from scratch. In most cases, prior knowledge cannot be directly copied from the source due to the different internal learning representations (e.g., from a human to a virtual agent). We need to re-summarize the usable transferred policy by observing the state-action behavior of the source agent. We propose methods that could analyze the offline confidence in summarized knowledge, which would help the agent better understand the source policy. Given the limited observation data, the agent can estimate which parts of the policy are trustworthy based on the confidence level. However, only measuring the offline confidence based on observed data is not enough. This dissertation extends the confidence analysis to the online learning setting, by further updating the confidence values based on how well the prior policies could adapt to the target tasks. Our online confidence mechanism could also remove manually defined sensitive parameters tuning and let the learner effectively adjust by itself. Confidence measurement could be done by leveraging supervised learning techniques but the demonstrations data could contain more sequential information like spatial or temporal structure. In addition to the confidence measurement of supervised methods, we further explore model-based methods to summarize the policy. Knowledge transfer is more challenging in real-life tasks, this dissertation would discuss techniques that enables the transfer from a model-based simulation to real robots. For extremely complicated learning areas like deep learning (DL), we propose correlated-feature networks that enable the agent to progressively reuse weights considering the feature relations between source and target tasks.
    URI
    http://hdl.handle.net/2376/17895
    Collections
    • Electronic Dissertations and Theses - Electrical Engineering and Computer Science
    • Electronic Dissertations

    Related items

    Showing items related by title, author, creator and subject.

    • The learning environment as place : an analysis of the United States Department of Education's six design principles for learning environments 

      Fritz, Catherine Mary, 1957- (2003)
    • ONLINE TEACHER EDUCATION: LEARNING COMMUNITY DEVELOPMENT AND SELF-DIRECTED PROFESSIONAL LEARNING 

      Herman, David (2019)
      With the rapid increase in the amount of online learning opportunities available has come with it an increase in challenges concerning how to support effective teaching and learning in online learning spaces. Through two ...
    • Learning from Human Teachers: Supporting How People Want to Teach in Interactive Machine Learning 

      Peng, Bei (2018)
      As the number of deployed robots grows, there will be an increasing need for humans to teach robots new skills that were not pre-programmed, without requiring these users to have any experience with programming or artificial ...