Jump to content

AIL722: Difference between revisions

From IITD Wiki
[checked revision][checked revision]
Creating course page via bot
 
Bot: wrap bare course codes in wikilinks
 
Line 4: Line 4:
| credits = 3
| credits = 3
| credit_structure = 3-0-0
| credit_structure = 3-0-0
| pre_requisites = Any one of ELL409 / ELL784 / AIL701 / AIL721
| pre_requisites = Any one of [[ELL409]] / [[ELL784]] / [[AIL701]] / [[AIL721]]
| overlaps = ELL729 approx. 50 %. [COL 333, COL341,
| overlaps = [[ELL729]] approx. 50 %. [COL 333, [[COL341]],
}}
}}


== AIL722 : Reinforcement Learning ==
== AIL722 : Reinforcement Learning ==
/ COL 341 / COL774 COL770, COL774, ELL409, ELL 784, ELL802, ELL888) < 10% Introduction and Basics of RL, Markov Decision Processes (MDPs), Dynamic Programming, Monte Carlo Methods (Prediction), Temporal difference Methods (Prediction), Monte Carlo, TD Method (Control), Monte Carlo, TD Method (Control - cont.), N-step TD, EligibiJity Traces, Model based RL, (Action-)Value Function Approximation, Value Function Approximation, Policy Gradient, Policy Gradient, Misc. Topics.
/ COL 341 / [[COL774]] [[COL770]], [[COL774]], [[ELL409]], ELL 784, [[ELL802]], [[ELL888]]) < 10% Introduction and Basics of RL, Markov Decision Processes (MDPs), Dynamic Programming, Monte Carlo Methods (Prediction), Temporal difference Methods (Prediction), Monte Carlo, TD Method (Control), Monte Carlo, TD Method (Control - cont.), N-step TD, EligibiJity Traces, Model based RL, (Action-)Value Function Approximation, Value Function Approximation, Policy Gradient, Policy Gradient, Misc. Topics.

Latest revision as of 16:21, 14 April 2026

AIL722
Reinforcement Learning
Credits 3
Structure 3-0-0
Pre-requisites Any one of ELL409 / ELL784 / AIL701 / AIL721
Overlaps ELL729 approx. 50 %. [COL 333, COL341,

AIL722 : Reinforcement Learning

/ COL 341 / COL774 COL770, COL774, ELL409, ELL 784, ELL802, ELL888) < 10% Introduction and Basics of RL, Markov Decision Processes (MDPs), Dynamic Programming, Monte Carlo Methods (Prediction), Temporal difference Methods (Prediction), Monte Carlo, TD Method (Control), Monte Carlo, TD Method (Control - cont.), N-step TD, EligibiJity Traces, Model based RL, (Action-)Value Function Approximation, Value Function Approximation, Policy Gradient, Policy Gradient, Misc. Topics.