QPICAR Deep Learning

dc.contributor.advisor Tuna Çakar
dc.contributor.author Beğde, Özge
dc.date.accessioned 2021-12-14T11:21:14Z
dc.date.available 2021-12-14T11:21:14Z
dc.date.issued 2021
dc.description.abstract The aim of the project is to train a smart tool kit named "Sunfounder Raspberry Pi Robot Car" to move without hitting the walls in a closed area. The goal is to maximize the driving time without crashing by reducing the number of hits. Ultrasonic sensor data collected from the vehicle are processed with reinforcement learning and deep reinforcement learning algorithms and the results are compared. In this study, Python programminglanguage is used. In this study, firstly, the Q-Learning method, which is a reinforcement learning algorithm based on Markov decision processes, is used. The method basically relies on a memory table, Q-Table, in which the Q-values of the agent moving from one state to another are kept. This table is updated according to the results of the Bellman equation in every action of the agent, and as a result of this iterative process, it is optimized to provide that the agent moves to maximize its rewards. Deep Q-Learning (DQN) is used as a deep reinforcement learning algorithm. This algorithm was developed by the DeepMind Technologies team in 2013. Basically, it is based on the use of the Bellman equation, which is an element of the Q-Learning method, incombination with neural networks. This method is often used for training agents in complex and multidimensional environments such as video games. Due to the different type of the data used on the algorithm, minor changes were made to adapt it to the study. RElu and Softplus are used as activation functions. The results of the training process show that the DQN algorithm has an important advantage in terms of training the agent in a short time. At this point, the results are in accordance with other academic studies demonstrating the success of the DQN algorithm for complex environments.For future work, by differentiating the equipment that collects data on the vehicle, different data types such as image, temperature value, oxygen value can be collected and processed. At the same time, with changes to the reward setup in the algorithm, the agent can be trained to move to a specific target or to take actions to avoid a specific target.
dc.identifier.citation Beğde, Ö. (2021). QPICAR Deep Learning. MEF Üniversitesi Fen Bilimleri Enstitüsü, Büyük Veri Analitiği Yüksek Lisans Programı. ss. 1-25
dc.identifier.uri https://hdl.handle.net/20.500.11779/1703
dc.language.iso en
dc.publisher MEF Üniversitesi Fen Bilimleri Enstitüsü
dc.rights info:eu-repo/semantics/openAccess
dc.subject Machine Learning, Reinforcement Learning, Q-Learning, Deep Reinforcement Learning
dc.title QPICAR Deep Learning
dc.title.alternative QPICAR derin öğrenme
dc.type Master's Degree Project
dspace.entity.type Publication
gdc.author.institutional Beğde, Özge
gdc.author.institutional Çakar, Tuna
gdc.coar.access open access
gdc.coar.type text::thesis::master thesis
gdc.description.department Lisansüstü Eğitim Enstitüsü, Büyük Veri Analitiği Yüksek Lisans Programı
gdc.description.publicationcategory YL-Bitirme Projesi
gdc.description.scopusquality N/A
gdc.description.startpage 1-25
gdc.description.wosquality N/A
relation.isAuthorOfPublication 10f8ce3b-94c2-40f0-9381-0725723768fe
relation.isAuthorOfPublication.latestForDiscovery 10f8ce3b-94c2-40f0-9381-0725723768fe
relation.isOrgUnitOfPublication 05ffa8cd-2a88-4676-8d3b-fc30eba0b7f3
relation.isOrgUnitOfPublication 0d54cd31-4133-46d5-b5cc-280b2c077ac3
relation.isOrgUnitOfPublication a6e60d5c-b0c7-474a-b49b-284dc710c078
relation.isOrgUnitOfPublication.latestForDiscovery 05ffa8cd-2a88-4676-8d3b-fc30eba0b7f3

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
FBE_BüyükVeriAnalitiği_ÖzgeBeğde.pdf
Size:
2.24 MB
Format:
Adobe Portable Document Format
Description:
YL-Proje Dosyası

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
0 B
Format:
Item-specific license agreed upon to submission
Description: