Proximal Policy Optimization for lane following

dc.contributor.advisorThill, S.
dc.contributor.authorGeurtjens, R. P.
dc.date.issued2020-07-10
dc.description.abstractThis thesis aimed to apply a state of the art reinforcement learning algorithm named proximal policy optimization on a complicated task with real world applicability in which sensor data is not always reliable. This algorithm was tested on the task of lane following. In order to do this the autonomous car simulator Carla was used. Semantic segmentation and Canny lter were discussed as methods to extract the lanes from the RGB sensor that the Carla simulator provided. The agent's performance was then examined on one of Carla's maps. In the end it turned out to be impossible to run the experiment due through hardware limitations. As an alternative, the algorithm was tested on the Luna lander environment, a game in which the agent had to land a rocket on the moon. Adding Gaussian noise to the agent's sensors did not prevent the algorithm from converging. It could be concluded from this that proximal policy optimization can derive an optimal policy on easy environments even if the sensor data is not completely reliable. There are, however, limits to the amount of noise that can be added.en_US
dc.embargo.lift10000-01-01
dc.embargo.typePermanent embargoen_US
dc.identifier.urihttps://theses.ubn.ru.nl/handle/123456789/12666
dc.language.isoenen_US
dc.thesis.facultyFaculteit der Sociale Wetenschappenen_US
dc.thesis.specialisationBachelor Artificial Intelligenceen_US
dc.thesis.studyprogrammeArtificial Intelligenceen_US
dc.thesis.typeBacheloren_US
dc.titleProximal Policy Optimization for lane followingen_US
Files
Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
1006223 Geurtjens.pdf
Size:
1.47 MB
Format:
Adobe Portable Document Format
Description: