From 347d4b2ad341248ab498e7c4e60fd2d41c328bdd Mon Sep 17 00:00:00 2001 From: =?utf8?q?J=C3=A9r=C3=B4me=20Benoit?= Date: Sun, 23 Feb 2025 23:05:04 +0100 Subject: [PATCH] refactor(reforcexy): refine configuration template MIME-Version: 1.0 Content-Type: text/plain; charset=utf8 Content-Transfer-Encoding: 8bit Signed-off-by: Jérôme Benoit --- ReforceXY/user_data/config-template.json | 2 +- ReforceXY/user_data/freqaimodels/ReforceXY.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/ReforceXY/user_data/config-template.json b/ReforceXY/user_data/config-template.json index 994fb94..38438cc 100644 --- a/ReforceXY/user_data/config-template.json +++ b/ReforceXY/user_data/config-template.json @@ -162,7 +162,7 @@ "profit_aim": 0.025, "win_reward_factor": 2 }, - "train_cycles": 25, + "train_cycles": 250, "add_state_info": true, "cpu_count": 6, "max_training_drawdown_pct": 0.02, diff --git a/ReforceXY/user_data/freqaimodels/ReforceXY.py b/ReforceXY/user_data/freqaimodels/ReforceXY.py index d7459eb..46d94bc 100644 --- a/ReforceXY/user_data/freqaimodels/ReforceXY.py +++ b/ReforceXY/user_data/freqaimodels/ReforceXY.py @@ -732,7 +732,7 @@ class ReforceXY(BaseReinforcementLearningModel): ) return factor - def calculate_reward(self, action) -> float: + def calculate_reward(self, action: int) -> float: """ An example reward function. This is the one function that users will likely wish to inject their own creativity into. -- 2.43.0