Please use this identifier to cite or link to this item: https://hdl.handle.net/20.500.11779/1703
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorTuna Çakar-
dc.contributor.authorBeğde, Özge-
dc.date.accessioned2021-12-14T11:21:14Z
dc.date.available2021-12-14T11:21:14Z
dc.date.issued2021-
dc.identifier.citationBeğde, Ö. (2021). QPICAR Deep Learning. MEF Üniversitesi Fen Bilimleri Enstitüsü, Büyük Veri Analitiği Yüksek Lisans Programı. ss. 1-25en_US
dc.identifier.urihttps://hdl.handle.net/20.500.11779/1703-
dc.description.abstractThe aim of the project is to train a smart tool kit named "Sunfounder Raspberry Pi Robot Car" to move without hitting the walls in a closed area. The goal is to maximize the driving time without crashing by reducing the number of hits. Ultrasonic sensor data collected from the vehicle are processed with reinforcement learning and deep reinforcement learning algorithms and the results are compared. In this study, Python programming language is used. In this study, firstly, the Q-Learning method, which is a reinforcement learning algorithm based on Markov decision processes, is used. The method basically relies on a memory table, Q-Table, in which the Q-values of the agent moving from one state to another are kept. This table is updated according to the results of the Bellman equation in every action of the agent, and as a result of this iterative process, it is optimized to provide that the agent moves to maximize its rewards. Deep Q-Learning (DQN) is used as a deep reinforcement learning algorithm. This algorithm was developed by the DeepMind Technologies team in 2013. Basically, it is based on the use of the Bellman equation, which is an element of the Q-Learning method, in combination with neural networks. This method is often used for training agents in complex and multidimensional environments such as video games. Due to the different type of the data used on the algorithm, minor changes were made to adapt it to the study. RElu and Softplus are used as activation functions. The results of the training process show that the DQN algorithm has an important advantage in terms of training the agent in a short time. At this point, the results are in accordance with other academic studies demonstrating the success of the DQN algorithm for complex environments. For future work, by differentiating the equipment that collects data on the vehicle, different data types such as image, temperature value, oxygen value can be collected and processed. At the same time, with changes to the reward setup in the algorithm, the agent can be trained to move to a specific target or to take actions to avoid a specific target.en_US
dc.language.isoenen_US
dc.publisherMEF Üniversitesi Fen Bilimleri Enstitüsüen_US
dc.rightsinfo:eu-repo/semantics/openAccessen_US
dc.subjectMachine Learning, Reinforcement Learning, Q-Learning, Deep Reinforcement Learningen_US
dc.titleQPICAR deep learningen_US
dc.title.alternativeQPICAR derin öğrenmeen_US
dc.typeMaster's Degree Projecten_US
dc.relation.publicationcategoryYL-Bitirme Projesien_US
dc.identifier.startpage1-25en_US
dc.departmentBüyük Veri Analitiği Yüksek Lisans Programıen_US
dc.institutionauthorBeğde, Özge-
item.openairecristypehttp://purl.org/coar/resource_type/c_18cf-
item.grantfulltextopen-
item.languageiso639-1en-
item.cerifentitytypePublications-
item.fulltextWith Fulltext-
item.openairetypeMaster's Degree Project-
Appears in Collections:FBE, Yüksek Lisans, Proje Koleksiyonu
Files in This Item:
File Description SizeFormat 
FBE_BüyükVeriAnalitiği_ÖzgeBeğde.pdfYL-Proje Dosyası2.29 MBAdobe PDFThumbnail
View/Open
Show simple item record



CORE Recommender

Page view(s)

2
checked on Jun 26, 2024

Google ScholarTM

Check





Items in GCRIS Repository are protected by copyright, with all rights reserved, unless otherwise indicated.