Penerapan Model Pembelajaran dengan Metode Reinforcement Learning Menggunakan Simulator Carla
DOI:
https://doi.org/10.30865/mib.v5i4.3169Keywords:
Self-driving Car, Reinforcement Learning, Convolutional Neural Network (CNN), Carla Car SimulatorAbstract
Artificial Intelligence is the study of how to make machines or computer programs have the intelligence or ability to do things that humans can do. The application of AI is currently in various ways, one of which is for self-driving cars. To be able to do a self-driving car, the AI that is implanted in a car must applied to the method to be able to walk on its path and be able to adapt to its environment. Reinforcement learning is one type of machine learning where agents learn something by doing certain actions and the results of those actions and try to maximize the gifts received through interactions with the environment that are reward negative or positive. In this research, we applied of the reinforcement learning method on the Carla Car simulator. The simulator is used to collect data using an RGB sensor, then modeling experiments which produce several models to be used in simulation experiments. The model is obtained by using the Convolutional Neural Network (CNN) algorithm with the NVIDIA architectural model. From the results of research based on experiments conducted obtained the best model obtained from the experimental model by comparing the maximum reward value, high accuracy and low loss is model 1 in the experimental model A with 100 episodes and model 4 in model B experiment with 150 episodesReferences
Abdelgawad, K., Gausemeier, J., Dumitrescu, R., & Grafe, M, "Networked Driving Simulation: Applications, State of the Art, and Design Considerations", MDPI, 2017.
A. Sallab, M. Abdou, E. Perot and S. Yogamani, "Deep reinforcement learning framework for autonomous driving", Electron. Imag., vol. 2017, no. 19, pp. 70-76, 2017.
Andreas, C., Muller, & Sarah, G, "Introduction to Machine Learning with Python (1st ed.)", O'Reilly Media, Inc., 2016.
C. K. Toth, Z. K, "New Source of Geospatial Data: Crowdsensing By Assisted and Autonomous Vehicle Technologies", The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, 2018.
Duan, Y., Andrychowicz, M., Stadie, B., Schneider, J., Sutskever, I., Abbeel, P., & Zaremba, W., "One Shot Imitation Learning," 2017.
Goldstone, W., "Unity 3.x Game Development Essentials", United Kingdom: Packt Publishing Ltd., 2011.
He, H., III, H. D., & Eisner, J., "Imitation Learning by Coaching," NIPS, 2012.
Kulic, R., & Vukic, Z., "Autonomous Vehicle Obstacle Avoiding and Goal Position Reaching by Virtual Obstacle", 2484-2491, 2017.
Lipton, Z. C., & Elkan, C., "Playing the Imitation Game with deep learning," 2016.
NVIDIA Corporation. "End to End Learning for Self-Driving Cars". arXiv, 1-9. 2016
Oh, J., Guo, Y., Singh, S., & Lee, H., "Self-Imitation Learning," Semantic Scholar. 2018.
Prasanna, B., Michael , S., Thomas , U., Venkat , V., & Stefan , W. "DeepHyper: Asynchronous Hyperparameter Search for Deep Neural Networks. High Performance Computing", 2018
S. Sharma, G. Tewolde and J. Kwon, "Behavioral cloning for lateral motion control of autonomous vehicles using deep learning", Proc. IEEE Int. Conf. Electro/Inf. Technol. (EIT), pp. 228-233, 2018.
Schroecker, Y., & Isbell, C. "State Aware Imitation Learning. Neural Information Processing Systems", pp. 1-10, USA: CA, 2017
Stuart, R., & Norvig, P., "Artificial Intelligence A Modern Approach (4th ed)", Pearson Education, Inc., 2021.
V. V. Dixit, S. Chand and D. J. Nair, "Autonomous vehicles: disengagements accidents and reaction times", PLoS one, vol. 11, no. 12, pp. e0168054, 2016.
Xiaoyong, Y., Pan, H., Qile, Z., & Xiaolin, L. "Adverserial Example: Attacks and Defenses for Deep Learning", 2019.
Zhang, X., Chen, M., & Zhan, X. "Reinforcement Learning for Driverless Cars using Transfer Learning", IEEE, 1069-1073. 2018
Downloads
Published
Issue
Section
License

This work is licensed under a Creative Commons Attribution 4.0 International License
Authors who publish with this journal agree to the following terms:
- Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under Creative Commons Attribution 4.0 International License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (Refer to The Effect of Open Access).