Skip to content

Summer Research - Tamer, Reinforcement Learning

Notifications You must be signed in to change notification settings

amathsow/TAMER-Summer-2018-Lab

 
 

Repository files navigation

Week 3

  • Berkeley Grid World
    • Setup TAMER on berkely gridworld
    • Make it able to save Tamer (Q table, log) on every trial
  • Experiment Creater and Resumer
    • Modify the original experiment launcher and resumer
    • Redirect all the output (print) to log files and make a copy of related files

Week 4

  • Reading Note
  • Berkeley Grid World
    • Add Evaluation metric: number of episode/step to reach optimal policy
    • Add instant feedback: pause and wait until human feedback
    • Random starting initial starting position
    • Softmax action selection for exploration with overflow protection

Week 5

  • Reading Note
  • Berkeley Grid World
    • Make environment deterministic
    • Value Iteration Experiment: use value iteration to find the optimal (converged) values (Q-Values) in the Gridworld
    • QValue Saver and Loader: save the computed q-values to json files, as well as read q-values from json files
    • Record and display policy agreement ratio
    • Add temperature control into Softmax action selection
    • Visualize last state and action
    • Statistics Module: visualize and compare experiment results

Week 6

Week 7

Week 8

Week 9

Week 10

  • Berkeley Grid World
    • Add supports for preferences-based agents

About

Summer Research - Tamer, Reinforcement Learning

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 86.4%
  • HTML 13.6%