[FreeCoursesOnline.Me] Coursera - Practical Reinforcement Learning

Torrent Details

[FreeCoursesOnline.Me] Coursera - Practical Reinforcement Learning[FreeCoursesOnline.Me] Coursera - Practical Reinforcement Learning

NAME
[FreeCoursesOnline.Me] Coursera - Practical Reinforcement Learning.torrent
CATEGORY
Other
INFOHASH
31b47a1285df93a33f1c80a563fd43b322fc434d
SIZE
1.4 GB in 99 files
ADDED
Uploaded on 03-10-2018 by our crawler pet called "Spidey".
SWARM
0 seeders & 0 peers
RATING
No votes yet.

Please login to vote for this torrent.


Description

[COURSERA] PRACTICAL REINFORCEMENT LEARNING [FCO]

About this course: Welcome to the Reinforcement Learning course. Here you will find out about: – foundations of RL methods: value/policy iteration, q-learning, policy gradient, etc. — with math & batteries included – using deep neural networks for RL tasks — also known as “the hype train” – state of the art RL algorithms — and how to apply duct tape to them for practical problems. – and, of course, teaching your neural network to play games — because that’s what everyone thinks RL is about. We’ll also use it for seq2seq and contextual bandits. Jump in. It’s gonna be fun!

For more Coursera and other Courses >>> https://www.freecoursesonline.me/
For More Udemy Free Courses >>> http://www.freetutorials.us

Discussion

Comments 0

Post Your Comment

Files in this torrent

FILENAMESIZE
001.Welcome/001. Why should you care.mp432.4 MB
001.Welcome/001. Why should you care.srt15.4 KB
001.Welcome/002. Reinforcement learning vs all.mp410.8 MB
001.Welcome/002. Reinforcement learning vs all.srt4.9 KB
002.Reinforcement Learning/003. Multi-armed bandit.mp417.9 MB
002.Reinforcement Learning/003. Multi-armed bandit.srt7.3 KB
002.Reinforcement Learning/004. Decision process & applications.mp423 MB
002.Reinforcement Learning/004. Decision process & applications.srt9.7 KB
003.Black box optimization/005. Markov Decision Process.mp418 MB
003.Black box optimization/005. Markov Decision Process.srt8.3 KB
003.Black box optimization/006. Crossentropy method.mp436 MB
003.Black box optimization/006. Crossentropy method.srt15.5 KB
003.Black box optimization/007. Approximate crossentropy method.mp419.3 MB
003.Black box optimization/007. Approximate crossentropy method.srt8.2 KB
003.Black box optimization/008. More on approximate crossentropy method.mp422.9 MB
003.Black box optimization/008. More on approximate crossentropy method.srt10.5 KB
004.All the cool stuff that isn't in the base track/009. Evolution strategies core idea.mp420.9 MB
004.All the cool stuff that isn't in the base track/009. Evolution strategies core idea.srt7.3 KB
004.All the cool stuff that isn't in the base track/010. Evolution strategies math problems.mp417.7 MB
004.All the cool stuff that isn't in the base track/010. Evolution strategies math problems.srt8.6 KB
004.All the cool stuff that isn't in the base track/011. Evolution strategies log-derivative trick.mp427.8 MB
004.All the cool stuff that isn't in the base track/011. Evolution strategies log-derivative trick.srt12.6 KB
004.All the cool stuff that isn't in the base track/012. Evolution strategies duct tape.mp421.2 MB
004.All the cool stuff that isn't in the base track/012. Evolution strategies duct tape.srt9.7 KB
004.All the cool stuff that isn't in the base track/013. Blackbox optimization drawbacks.mp415.2 MB
004.All the cool stuff that isn't in the base track/013. Blackbox optimization drawbacks.srt7.3 KB
005.Striving for reward/014. Reward design.mp449.7 MB
005.Striving for reward/014. Reward design.srt23.2 KB
006.Bellman equations/015. State and Action Value Functions.mp437.3 MB
006.Bellman equations/015. State and Action Value Functions.srt18.2 KB
006.Bellman equations/016. Measuring Policy Optimality.mp418.1 MB
006.Bellman equations/016. Measuring Policy Optimality.srt8.5 KB
007.Generalized Policy Iteration/017. Policy evaluation & improvement.mp431.9 MB
007.Generalized Policy Iteration/017. Policy evaluation & improvement.srt14.5 KB
007.Generalized Policy Iteration/018. Policy and value iteration.mp424.2 MB
007.Generalized Policy Iteration/018. Policy and value iteration.srt12.1 KB
008.Model-free learning/019. Model-based vs model-free.mp428.8 MB
008.Model-free learning/019. Model-based vs model-free.srt14.1 KB
008.Model-free learning/020. Monte-Carlo & Temporal Difference; Q-learning.mp430.1 MB
008.Model-free learning/020. Monte-Carlo & Temporal Difference; Q-learning.srt14.5 KB
008.Model-free learning/021. Exploration vs Exploitation.mp428.2 MB
008.Model-free learning/021. Exploration vs Exploitation.srt14 KB
008.Model-free learning/022. Footnote Monte-Carlo vs Temporal Difference.mp410.3 MB
008.Model-free learning/022. Footnote Monte-Carlo vs Temporal Difference.srt4.8 KB
009.On-policy vs off-policy/023. Accounting for exploration. Expected Value SARSA..mp437.7 MB
009.On-policy vs off-policy/023. Accounting for exploration. Expected Value SARSA..srt17.3 KB
010.Experience Replay/024. On-policy vs off-policy; Experience replay.mp426.7 MB
010.Experience Replay/024. On-policy vs off-policy; Experience replay.srt11.2 KB
011.Limitations of Tabular Methods/025. Supervised & Reinforcement Learning.mp450.6 MB
011.Limitations of Tabular Methods/025. Supervised & Reinforcement Learning.srt25.4 KB
011.Limitations of Tabular Methods/026. Loss functions in value based RL.mp433.8 MB
011.Limitations of Tabular Methods/026. Loss functions in value based RL.srt15.2 KB
011.Limitations of Tabular Methods/027. Difficulties with Approximate Methods.mp447 MB
011.Limitations of Tabular Methods/027. Difficulties with Approximate Methods.srt21.9 KB
012.Case Study Deep Q-Network/028. DQN bird's eye view.mp427.8 MB
012.Case Study Deep Q-Network/028. DQN bird's eye view.srt11.4 KB
012.Case Study Deep Q-Network/029. DQN the internals.mp429.6 MB
012.Case Study Deep Q-Network/029. DQN the internals.srt12.3 KB
013.Honor/030. DQN statistical issues.mp419.2 MB
013.Honor/030. DQN statistical issues.srt9.2 KB
013.Honor/031. Double Q-learning.mp420.5 MB
013.Honor/031. Double Q-learning.srt9.4 KB
013.Honor/032. More DQN tricks.mp433.9 MB
013.Honor/032. More DQN tricks.srt16.4 KB
013.Honor/033. Partial observability.mp457.2 MB
013.Honor/033. Partial observability.srt27.7 KB
014.Policy-based RL vs Value-based RL/034. Intuition.mp434.9 MB
014.Policy-based RL vs Value-based RL/034. Intuition.srt15.6 KB
014.Policy-based RL vs Value-based RL/035. All Kinds of Policies.mp416 MB
014.Policy-based RL vs Value-based RL/035. All Kinds of Policies.srt7.4 KB
014.Policy-based RL vs Value-based RL/036. Policy gradient formalism.mp431.6 MB
014.Policy-based RL vs Value-based RL/036. Policy gradient formalism.srt13.3 KB
014.Policy-based RL vs Value-based RL/037. The log-derivative trick.mp413.3 MB
014.Policy-based RL vs Value-based RL/037. The log-derivative trick.srt5.9 KB
015.REINFORCE/038. REINFORCE.mp431.4 MB
015.REINFORCE/038. REINFORCE.srt14 KB
016.Actor-critic/039. Advantage actor-critic.mp424.6 MB
016.Actor-critic/039. Advantage actor-critic.srt11.8 KB
016.Actor-critic/040. Duct tape zone.mp417.5 MB
016.Actor-critic/040. Duct tape zone.srt7.8 KB
016.Actor-critic/041. Policy-based vs Value-based.mp416.8 MB
016.Actor-critic/041. Policy-based vs Value-based.srt7.1 KB
016.Actor-critic/042. Case study A3C.mp426.1 MB
016.Actor-critic/042. Case study A3C.srt11.1 KB
016.Actor-critic/043. A3C case study (2 2).mp415 MB
016.Actor-critic/043. A3C case study (2 2).srt6 KB
016.Actor-critic/044. Combining supervised & reinforcement learning.mp424 MB
016.Actor-critic/044. Combining supervised & reinforcement learning.srt11.9 KB
017.Measuting exploration/045. Recap bandits.mp424.7 MB
017.Measuting exploration/045. Recap bandits.srt11.9 KB
017.Measuting exploration/046. Regret measuring the quality of exploration.mp421.3 MB
017.Measuting exploration/046. Regret measuring the quality of exploration.srt10.2 KB
017.Measuting exploration/047. The message just repeats. 'Regret, Regret, Regret.'.mp418.4 MB
017.Measuting exploration/047. The message just repeats. 'Regret, Regret, Regret.'.srt8.7 KB
018.Uncertainty-based exploration/048. Intuitive explanation.mp422.3 MB
018.Uncertainty-based exploration/048. Intuitive explanation.srt10.9 KB
018.Uncertainty-based exploration/049. Thompson Sampling.mp417.1 MB
018.Uncertainty-based exploration/049. Thompson Sampling.srt7.9 KB
018.Uncertainty-based exploration/050. Optimism in face of uncertainty.mp416.5 MB

Alternative Torrents for 'FreeCoursesOnline.Me Coursera Practical Reinforcement Learning'.

There are no alternative torrents found.