-
Notifications
You must be signed in to change notification settings - Fork 6
Using
Ngô Xuân Phong edited this page May 17, 2023
·
5 revisions
* Agent(state, agent_data):
input: state, data
output: action(int), data
* getValidActions(state):
input: state(np.float64)
output: np.array valid action in turn of Environment
* getReward(state):
input: state(np.float64)
output: -1, 0, 1
* getActionSize():
input: None
output: int, count of array action size of Env
* getStateSize():
input: None
output: int, count of array state size of Env
from numba import njit
import numpy as np
@njit()
def Agent(state, agent_data):
validActions = env.getValidActions(state)
actions = np.where(validActions==1)[0]
action = np.random.choice(actions)
return arr_action[idx], agent_data
from setup import make
env = make('SushiGo')
More env please read Environments
env = make('SushiGo)
count_win, agent_data = env.numba_main_2(Agent, count_game_train = 1, agent_data = [0], level = 0)
Var | Type | Description |
---|---|---|
count_game_train | int | matches of a environment |
agent_data | any | data train of agent |
level | 0, 1, -1 | level of environment (update more later) |
env.render(Agent=Agent, per_data=[0], level=0, max_temp_frame=100)
You may be interested in FAQ.
Contributions are Welcome!
Contributions are Welcome!