Training Flappy Bird with Deep Q Network and SARSA
Downloads
Deep Q-Network (DQN) is implemented in this Flappy Bird Game, and the purpose of the project is to tweak and change the parameters to meet the desired outcome, which is passing pipes if the agent can. DQN are employed into the project to maximize cumulative reward while making decisions in real-time gameplay. In this project, the Flappy Bird game environment is set up along with the state and action spaces, the Q-network, an experience replay buffer, and a training loop. Using ε-greedy policies, the agent learns to make the best decisions by balancing exploration and exploitation. DeepMind’s experience replay was employed to enhance the stability of learning. Adam optimizer and SGD optimizer are monitoring the agent’s train/loss, and the α and ε will be modified to make a comparison between different parameters.
Downloads
Copyright (c) 2024 Journal of Applied Technology and Innovation

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.



