Skip to content

gt-coar/NesaraREU20

Repository files navigation

NesaraREU20

Nesara's REU 2020 Project Code Repository

Contents

  • Plot J
    • Contains the implementation of 3D visualization of the objective function of a 2 State 2 Action MDP
  • Policy and Value Iteration
    • Policy Iteration
    • Value Iteration
    • Policy vs Value Iteration (Comparison)
  • Direct Parameterization
    • Contains the implementation of Direct Policy Parameterization
  • SoftMax Parameterization
    • Contains the implementation of SoftMax Policy Parameterization
  • Natural actor critic
    • Contains the implementation of the paper : "Finite Sample Analysis of Two-Time-Scale NaturalActor-Critic Algorithm"
  • Natural Policy Gradient : MDP Case
    • Natural Policy Gradient with Softmax Parameterization
    • Natural Policy Gradient with Softmax Parameterization : Function Approximation
  • Mirror and Lazy Mirror Descent
    • Mirror Descent
    • Lazy Mirror Descent
  • TRPO
    • Contains the implementation of Trust Region Policy Optimization Algorithm
  • Compare Algorithms : Notebooks containing comparisons between various algorithms implemented in this repository
    • Compare time and iteration : Compares the total time and total iterations taken by the following algorithms to converge in separate plots
      • Policy Iteration
      • Value Iteration
      • Policy Gradient with Direct Parameterization (Constant step-size)
      • Policy Gradient with Direct Parameterization (Time variant step-size)
      • Policy Gradient with Softmax Parameterization (Constant step-size)
      • Policy Gradient with Softmax Parameterization (Time variant step-size)
    • Compare mirror and lazy mirror descent : Compares the convergence between mirror and lazy mirror descent algorithms
    • Compare convergence of policy gradient algos : Compares the total iterations taken by the following policy gradient based algorithms to converge
      • Policy Gradient with Direct Parameterization (Constant step-size)
      • Policy Gradient with Softmax Parameterization (Constant step-size)
      • Policy Gradient with Softmax Parameterization (Constant step-size) : Function Approximation
      • Natural Policy Gradient with Softmax Parameterization (Constant step-size)
      • Natural Policy Gradient with Softmax Parameterization (Constant step-size) : Function Approximation

About

Nesara's REU 2020 Project Code

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published