forked from AI4Finance-Foundation/FinRL-Meta
-
Notifications
You must be signed in to change notification settings - Fork 0
/
train.py
99 lines (82 loc) · 3.23 KB
/
train.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
# use three drl libraries: elegantrl, rllib, and SB3
from agents.elegantrl_models import DRLAgent as DRLAgent_erl
from agents.rllib_models import DRLAgent as DRLAgent_rllib
from agents.stablebaselines3_models import DRLAgent as DRLAgent_sb3
from meta.data_processor import DataProcessor
def train(
start_date,
end_date,
ticker_list,
data_source,
time_interval,
technical_indicator_list,
drl_lib,
env,
model_name,
if_vix=True,
**kwargs
):
# process data using unified data processor
dp = DataProcessor(data_source, start_date, end_date, time_interval, **kwargs)
price_array, tech_array, turbulence_array = dp.run(
ticker_list, technical_indicator_list, if_vix
)
data_config = {
"price_array": price_array,
"tech_array": tech_array,
"turbulence_array": turbulence_array,
}
# build environment using processed data
env_instance = env(config=data_config)
# read parameters and load agents
cwd = kwargs.get("cwd", "./" + str(model_name)) # cwd: current_working_dir
if drl_lib == "elegantrl":
break_step = kwargs.get("break_step", 1e6) # total_training_steps
erl_params = kwargs.get("erl_params") # see notebooks for examples.
agent = DRLAgent_erl(
env=env,
price_array=price_array,
tech_array=tech_array,
turbulence_array=turbulence_array,
)
model = agent.get_model(model_name, model_kwargs=erl_params)
trained_model = agent.train_model(
model=model, cwd=cwd, total_timesteps=break_step
) # erl model is automated saved in cwd
elif drl_lib == "rllib":
total_episodes = kwargs.get(
"total_episodes", 100
) # rllib uses total training episodes instead of steps.
rllib_params = kwargs.get("rllib_params")
agent_rllib = DRLAgent_rllib(
env=env,
price_array=price_array,
tech_array=tech_array,
turbulence_array=turbulence_array,
)
model, model_config = agent_rllib.get_model(model_name)
model_config["lr"] = rllib_params["lr"] # learning_rate
model_config["train_batch_size"] = rllib_params["train_batch_size"]
model_config["gamma"] = rllib_params["gamma"]
trained_model = agent_rllib.train_model(
model=model,
model_name=model_name,
model_config=model_config,
total_episodes=total_episodes,
)
trained_model.save(cwd)
elif drl_lib == "stable_baselines3":
total_timesteps = kwargs.get("total_timesteps", 1e6)
agent_params = kwargs.get("agent_params")
agent = DRLAgent_sb3(env=env_instance)
model = agent.get_model(model_name, model_kwargs=agent_params)
trained_model = agent.train_model(
model=model,
tb_log_name=model_name,
total_timesteps=total_timesteps,
)
print("Training finished!")
trained_model.save(cwd)
print("Trained model saved in " + str(cwd))
else:
raise ValueError("DRL library input is NOT supported. Please check.")