Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings
Permalink

Comparing changes

Choose two branches to see what’s changed or to start a new pull request. If you need to, you can also or learn more about diff comparisons.

Open a pull request

Create a new pull request by comparing changes across two branches. If you need to, you can also . Learn more about diff comparisons here.
base repository: 8000net/LectureNotesMaster
Failed to load repositories. Confirm that selected base ref is valid, then try again.
Loading
base: master
Choose a base ref
Loading
...
head repository: 8000net/LectureNotesMaster
Failed to load repositories. Confirm that selected head ref is valid, then try again.
Loading
compare: RLAdvanced
Choose a head ref
Loading
Checking mergeability… Don’t worry, you can still create the pull request.
  • 1 commit
  • 1 file changed
  • 1 contributor

Commits on Aug 9, 2020

  1. Adds some extra notes on RL training

    I've recently been trying to take on RL again (bit of redemption after my
    glaring defeat to box2d car racing 2 years ago!).  In this pursuit I
    came across some new analogies that were extremely useful to me in
    creating a mental model for how some of these techniques work.  Thought
    they may be useful to some of your students.
    
    This commit
    - adds an analogy for the purpose of the target network
    - emphasizes the _reason_ experience replay works
    - adds a section on advanced RL techniques used to overcome sparse
    reward functions
    LukeWood committed Aug 9, 2020
    Configuration menu
    Copy the full SHA
    38d95fd View commit details
    Browse the repository at this point in the history
Loading
Morty Proxy This is a proxified and sanitized view of the page, visit original site.