This course is your perfect entry point into the exciting field of Reinforcement Learning where digital Artificial Intelligence agents are built to automatically learn how to make sequential decisions through trial–and–error. Specifically, this course focuses on the Multi–Armed Bandit problems and the practical hands–on implementation of various algorithmic strategies for balancing between exploration and exploitation. Whenever you desire to consistently make the best choice out of a limited number of options over time, you are dealing with a Multi–Armed Bandit problem and this course teaches you every detail you need to know to be able to build realistic business agents to handle such situations.
With very concise explanations, this course teaches you how to confidently translate seemingly scary mathematical formulas into Python code painlessly. We understand that not many of us are technically adept in the subject of mathematics so this course intentionally stays away from maths unless it is necessary. And even when it becomes necessary to talk about mathematics, the approach taken in this course is such that anyone with basic algebra skills can understand and most importantly easily translate the maths into code and build useful intuitions in the process.
Some of the algorithmic strategies taught in this course are Epsilon Greedy, Softmax Exploration, Optimistic Initialization, Upper Confidence Bounds, and Thompson Sampling. With these tools under your belt, you are adequately equipped to readily build and deploy AI agents that can handle critical business operations under uncertainties.
Specification: Practical Multi-Armed Bandit Algorithms in Python
|
User Reviews
Be the first to review “Practical Multi-Armed Bandit Algorithms in Python” Cancel reply
This site uses Akismet to reduce spam. Learn how your comment data is processed.
Price | $12.99 |
---|---|
Provider | |
Duration | 5.5 hours |
Year | 2022 |
Level | All |
Language | English ... |
Certificate | Yes |
Quizzes | Yes |
$19.99 $12.99
There are no reviews yet.