Multiarm Bandits: explore-then-exploit, Upper Confidence Bounds (UCB) Algorithms | Intro to RL

No views

Mihai Nica Lectures

40 minutes ago

Multiarm Bandits: explore-then-exploit, Upper Confidence Bounds (UCB) Algorithms | Intro to RL

Multiarm Bandits: explore-then-exploit, Upper Confidence Bounds (UCB) Algorithms | Intro to RL