×

Samuel Arzt's video: AI Learns Parallel Parking - Deep Reinforcement Learning

@AI Learns Parallel Parking - Deep Reinforcement Learning
Big thanks to Hostinger for sponsoring this video! Go to https://hostinger.com?REFERRALCODE=1SAMUEL08 and get 20% off your hosting plan. I'm really glad how my new portfolio worked out! https://samuelarzt.com Also check out the original parking video: https://youtu.be/VMp6pq6_QjI Two AI fight for the same parking spot: https://youtu.be/CqYKhbyHFtA Neural Networks explained in a Minute: https://youtu.be/rEDzUT3ymw4 Subscribe for more content like this: https://www.youtube.com/channel/UC_eerU4SleeptEbD2AA_nDw?sub_confirmation=1 Follow me on Twitter for more frequent updates on my projects: https://twitter.com/SamuelArzt Last time we trained an AI how to park (https://youtu.be/VMp6pq6_QjI). A lot of people suggested in the comments of that video to try parallel parking next. So that's what this video is all about. We are using the same methods as last time and try different adjustments to the learning algorithm and environment in order to make the agent more generalizing and precise. The simulation was implemented using Unity's ML-Agents framework (https://unity3d.com/machine-learning). The AI consists of a deep Neural Network with 3 hidden layers of 128 neurons each. It is trained with the Proximal Policy Optimization (PPO) algorithm, which is a Reinforcement Learning approach. Basically, the input of the Neural Network are the readings of eight depth sensors, the car's current speed and position, as well as its relative position to the target. The outputs of the Neural Network are interpreted as engine force, braking force and turning force. These outputs can be seen at the top right corner of the zoomed out camera shots. The AI starts off with random behaviour, i.e. the Neural Network is initialized with random weights. It then gradually learns to solve the task by reacting to environment feedback accordingly. The environment tells the AI whether it is doing good or bad with positive or negative reward signals. The training was done on a computer with an i5 (7th or 8th gen) and a GTX 1070 with 100x simulation speed, using 6 instances of the environment and up to 6 processes running in parallel. Music from Bensound.com: Timelapse Music: "The Elevator Bossa Nova" Outro: "All That" Background Music 1: Drops of H2O ( The Filtered Water Treatment ) by J.Lang (c) copyright 2012 Licensed under a Creative Commons Attribution (3.0) license. http://dig.ccmixter.org/files/djlang59/37792 Ft: Airtone

3.3K

184
Samuel Arzt
Subscribers
49.5K
Total Post
8
Total Views
14.6M
Avg. Views
1.8M
View Profile
This video was published on 2020-04-10 21:30:10 GMT by @Samuel-Arzt on Youtube. Samuel Arzt has total 49.5K subscribers on Youtube and has a total of 8 video.This video has received 3.3K Likes which are lower than the average likes that Samuel Arzt gets . @Samuel-Arzt receives an average views of 1.8M per video on Youtube.This video has received 184 comments which are lower than the average comments that Samuel Arzt gets . Overall the views for this video was lower than the average for the profile.Samuel Arzt #ArtificialIntelligence #MachineLearning #ReinforcementLearning #AI #NeuralNetworks #hostinger #inspeedwebelieve #speedfreak has been used frequently in this Post.

Other post by @Samuel Arzt