2018-03-02 13:46:32 +00:00
|
|
|
# pragma pylint: disable=too-many-instance-attributes, pointless-string-statement
|
2017-11-25 00:04:11 +00:00
|
|
|
|
2018-03-02 13:46:32 +00:00
|
|
|
"""
|
|
|
|
This module contains the hyperopt logic
|
|
|
|
"""
|
2017-11-25 00:04:11 +00:00
|
|
|
|
2019-11-06 18:33:15 +00:00
|
|
|
import locale
|
2017-11-25 01:04:37 +00:00
|
|
|
import logging
|
2018-01-23 14:56:12 +00:00
|
|
|
import sys
|
2019-08-15 18:39:04 +00:00
|
|
|
from collections import OrderedDict
|
2017-10-30 19:41:36 +00:00
|
|
|
from operator import itemgetter
|
2019-01-06 13:47:38 +00:00
|
|
|
from pathlib import Path
|
|
|
|
from pprint import pprint
|
2019-08-02 19:22:58 +00:00
|
|
|
from typing import Any, Dict, List, Optional
|
2017-10-19 14:12:49 +00:00
|
|
|
|
2019-08-15 18:39:04 +00:00
|
|
|
import rapidjson
|
2019-08-09 11:48:57 +00:00
|
|
|
from colorama import Fore, Style
|
2019-11-06 18:33:15 +00:00
|
|
|
from colorama import init as colorama_init
|
|
|
|
from joblib import (Parallel, cpu_count, delayed, dump, load,
|
|
|
|
wrap_non_picklable_objects)
|
2019-01-06 13:47:38 +00:00
|
|
|
from pandas import DataFrame
|
2018-06-19 06:09:54 +00:00
|
|
|
from skopt import Optimizer
|
2018-03-22 08:27:13 +00:00
|
|
|
from skopt.space import Dimension
|
2018-06-18 19:40:36 +00:00
|
|
|
|
2019-10-27 09:56:38 +00:00
|
|
|
from freqtrade.data.history import get_timeframe, trim_dataframe
|
2019-11-23 09:20:41 +00:00
|
|
|
from freqtrade.misc import plural, round_dict
|
2018-03-02 15:22:00 +00:00
|
|
|
from freqtrade.optimize.backtesting import Backtesting
|
2019-08-14 10:25:49 +00:00
|
|
|
# Import IHyperOpt and IHyperOptLoss to allow unpickling classes from these modules
|
|
|
|
from freqtrade.optimize.hyperopt_interface import IHyperOpt # noqa: F4
|
2019-07-17 05:14:27 +00:00
|
|
|
from freqtrade.optimize.hyperopt_loss_interface import IHyperOptLoss # noqa: F4
|
2019-11-06 18:33:15 +00:00
|
|
|
from freqtrade.resolvers.hyperopt_resolver import (HyperOptLossResolver,
|
|
|
|
HyperOptResolver)
|
2019-04-25 08:11:04 +00:00
|
|
|
|
2018-03-25 19:37:14 +00:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2019-04-25 08:11:04 +00:00
|
|
|
|
2019-05-10 07:54:44 +00:00
|
|
|
INITIAL_POINTS = 30
|
2019-09-23 08:59:34 +00:00
|
|
|
|
|
|
|
# Keep no more than 2*SKOPT_MODELS_MAX_NUM models
|
|
|
|
# in the skopt models list
|
|
|
|
SKOPT_MODELS_MAX_NUM = 10
|
|
|
|
|
2018-07-02 08:44:33 +00:00
|
|
|
MAX_LOSS = 100000 # just a big enough number to be bad result in loss optimization
|
|
|
|
|
2018-03-25 19:37:14 +00:00
|
|
|
|
2019-08-23 21:10:35 +00:00
|
|
|
class Hyperopt:
|
2018-01-23 14:56:12 +00:00
|
|
|
"""
|
2018-03-02 13:46:32 +00:00
|
|
|
Hyperopt class, this class contains all the logic to run a hyperopt simulation
|
2018-01-23 14:56:12 +00:00
|
|
|
|
2018-03-02 13:46:32 +00:00
|
|
|
To run a backtest:
|
|
|
|
hyperopt = Hyperopt(config)
|
|
|
|
hyperopt.start()
|
2018-01-23 14:56:12 +00:00
|
|
|
"""
|
2018-03-02 13:46:32 +00:00
|
|
|
def __init__(self, config: Dict[str, Any]) -> None:
|
2019-08-23 21:10:35 +00:00
|
|
|
self.config = config
|
|
|
|
|
2018-04-01 11:06:50 +00:00
|
|
|
self.custom_hyperopt = HyperOptResolver(self.config).hyperopt
|
2018-03-22 08:27:13 +00:00
|
|
|
|
2019-09-18 19:57:17 +00:00
|
|
|
self.backtesting = Backtesting(self.config)
|
|
|
|
|
2019-07-16 04:27:23 +00:00
|
|
|
self.custom_hyperoptloss = HyperOptLossResolver(self.config).hyperoptloss
|
|
|
|
self.calculate_loss = self.custom_hyperoptloss.hyperopt_loss_function
|
|
|
|
|
2019-07-31 05:07:46 +00:00
|
|
|
self.trials_file = (self.config['user_data_dir'] /
|
|
|
|
'hyperopt_results' / 'hyperopt_results.pickle')
|
|
|
|
self.tickerdata_pickle = (self.config['user_data_dir'] /
|
|
|
|
'hyperopt_results' / 'hyperopt_tickerdata.pkl')
|
2019-07-30 08:47:28 +00:00
|
|
|
self.total_epochs = config.get('epochs', 0)
|
2019-08-01 17:33:45 +00:00
|
|
|
|
2018-03-02 13:46:32 +00:00
|
|
|
self.current_best_loss = 100
|
|
|
|
|
2019-07-16 03:50:27 +00:00
|
|
|
if not self.config.get('hyperopt_continue'):
|
2019-07-15 18:17:15 +00:00
|
|
|
self.clean_hyperopt()
|
2019-07-16 03:50:27 +00:00
|
|
|
else:
|
|
|
|
logger.info("Continuing on previous hyperopt results.")
|
|
|
|
|
2018-06-22 10:02:26 +00:00
|
|
|
# Previous evaluations
|
2018-06-30 06:54:31 +00:00
|
|
|
self.trials: List = []
|
2018-03-02 13:46:32 +00:00
|
|
|
|
2019-11-23 08:32:33 +00:00
|
|
|
self.num_trials_saved = 0
|
|
|
|
|
2019-07-15 18:28:55 +00:00
|
|
|
# Populate functions here (hasattr is slow so should not be run during "regular" operations)
|
2019-09-16 18:22:07 +00:00
|
|
|
if hasattr(self.custom_hyperopt, 'populate_indicators'):
|
|
|
|
self.backtesting.strategy.advise_indicators = \
|
|
|
|
self.custom_hyperopt.populate_indicators # type: ignore
|
2019-07-15 18:28:55 +00:00
|
|
|
if hasattr(self.custom_hyperopt, 'populate_buy_trend'):
|
2019-09-18 19:57:17 +00:00
|
|
|
self.backtesting.strategy.advise_buy = \
|
|
|
|
self.custom_hyperopt.populate_buy_trend # type: ignore
|
2019-07-15 18:28:55 +00:00
|
|
|
if hasattr(self.custom_hyperopt, 'populate_sell_trend'):
|
2019-09-18 19:57:17 +00:00
|
|
|
self.backtesting.strategy.advise_sell = \
|
|
|
|
self.custom_hyperopt.populate_sell_trend # type: ignore
|
2019-07-15 18:28:55 +00:00
|
|
|
|
2019-08-02 19:22:58 +00:00
|
|
|
# Use max_open_trades for hyperopt as well, except --disable-max-market-positions is set
|
2019-07-15 18:28:55 +00:00
|
|
|
if self.config.get('use_max_market_positions', True):
|
|
|
|
self.max_open_trades = self.config['max_open_trades']
|
|
|
|
else:
|
|
|
|
logger.debug('Ignoring max_open_trades (--disable-max-market-positions was used) ...')
|
|
|
|
self.max_open_trades = 0
|
2019-09-25 00:41:22 +00:00
|
|
|
self.position_stacking = self.config.get('position_stacking', False)
|
2019-07-15 18:28:55 +00:00
|
|
|
|
2019-08-01 20:57:26 +00:00
|
|
|
if self.has_space('sell'):
|
2019-10-05 10:29:59 +00:00
|
|
|
# Make sure use_sell_signal is enabled
|
|
|
|
if 'ask_strategy' not in self.config:
|
|
|
|
self.config['ask_strategy'] = {}
|
|
|
|
self.config['ask_strategy']['use_sell_signal'] = True
|
2019-08-01 20:57:26 +00:00
|
|
|
|
2019-07-21 14:07:06 +00:00
|
|
|
@staticmethod
|
|
|
|
def get_lock_filename(config) -> str:
|
|
|
|
|
|
|
|
return str(config['user_data_dir'] / 'hyperopt.lock')
|
|
|
|
|
2019-07-15 18:17:15 +00:00
|
|
|
def clean_hyperopt(self):
|
|
|
|
"""
|
|
|
|
Remove hyperopt pickle files to restart hyperopt.
|
|
|
|
"""
|
2019-07-21 14:07:06 +00:00
|
|
|
for f in [self.tickerdata_pickle, self.trials_file]:
|
2019-07-15 18:17:15 +00:00
|
|
|
p = Path(f)
|
|
|
|
if p.is_file():
|
|
|
|
logger.info(f"Removing `{p}`.")
|
|
|
|
p.unlink()
|
|
|
|
|
2018-06-19 06:09:54 +00:00
|
|
|
def get_args(self, params):
|
2019-09-16 18:22:07 +00:00
|
|
|
|
|
|
|
dimensions = self.dimensions
|
|
|
|
|
2018-06-19 06:09:54 +00:00
|
|
|
# Ensure the number of dimensions match
|
|
|
|
# the number of parameters in the list x.
|
|
|
|
if len(params) != len(dimensions):
|
2018-07-03 08:17:41 +00:00
|
|
|
raise ValueError('Mismatch in number of search-space dimensions. '
|
|
|
|
f'len(dimensions)=={len(dimensions)} and len(x)=={len(params)}')
|
2018-06-19 06:09:54 +00:00
|
|
|
|
|
|
|
# Create a dict where the keys are the names of the dimensions
|
|
|
|
# and the values are taken from the list of parameters x.
|
|
|
|
arg_dict = {dim.name: value for dim, value in zip(dimensions, params)}
|
|
|
|
return arg_dict
|
|
|
|
|
2019-11-23 08:32:33 +00:00
|
|
|
def save_trials(self, final: bool = False) -> None:
|
2018-03-02 13:46:32 +00:00
|
|
|
"""
|
|
|
|
Save hyperopt trials to file
|
|
|
|
"""
|
2019-11-23 08:32:33 +00:00
|
|
|
num_trials = len(self.trials)
|
|
|
|
if num_trials > self.num_trials_saved:
|
2019-11-23 09:20:41 +00:00
|
|
|
logger.info(f"Saving {num_trials} {plural(num_trials, 'epoch')}.")
|
2018-07-03 19:51:48 +00:00
|
|
|
dump(self.trials, self.trials_file)
|
2019-11-23 08:32:33 +00:00
|
|
|
self.num_trials_saved = num_trials
|
|
|
|
if final:
|
2019-11-23 09:20:41 +00:00
|
|
|
logger.info(f"{num_trials} {plural(num_trials, 'epoch')} "
|
|
|
|
f"saved to '{self.trials_file}'.")
|
2018-03-02 13:46:32 +00:00
|
|
|
|
2018-06-22 10:02:26 +00:00
|
|
|
def read_trials(self) -> List:
|
2018-03-02 13:46:32 +00:00
|
|
|
"""
|
|
|
|
Read hyperopt trials file
|
|
|
|
"""
|
2019-08-25 18:38:51 +00:00
|
|
|
logger.info("Reading Trials from '%s'", self.trials_file)
|
2018-07-03 19:51:48 +00:00
|
|
|
trials = load(self.trials_file)
|
2019-07-21 13:56:44 +00:00
|
|
|
self.trials_file.unlink()
|
2018-03-02 13:46:32 +00:00
|
|
|
return trials
|
|
|
|
|
|
|
|
def log_trials_result(self) -> None:
|
|
|
|
"""
|
|
|
|
Display Best hyperopt result
|
|
|
|
"""
|
2019-11-23 08:32:33 +00:00
|
|
|
# This is printed when Ctrl+C is pressed quickly, before first epochs have
|
|
|
|
# a chance to be evaluated.
|
|
|
|
if not self.trials:
|
|
|
|
print("No epochs evaluated yet, no best result.")
|
|
|
|
return
|
|
|
|
|
2018-06-22 10:02:26 +00:00
|
|
|
results = sorted(self.trials, key=itemgetter('loss'))
|
|
|
|
best_result = results[0]
|
2019-08-01 20:57:26 +00:00
|
|
|
params = best_result['params']
|
2019-07-30 08:47:28 +00:00
|
|
|
log_str = self.format_results_logstring(best_result)
|
2019-08-03 08:05:05 +00:00
|
|
|
print(f"\nBest result:\n\n{log_str}\n")
|
2019-08-15 18:39:04 +00:00
|
|
|
|
|
|
|
if self.config.get('print_json'):
|
2019-08-15 20:13:46 +00:00
|
|
|
result_dict: Dict = {}
|
2019-08-15 18:39:04 +00:00
|
|
|
if self.has_space('buy') or self.has_space('sell'):
|
|
|
|
result_dict['params'] = {}
|
|
|
|
if self.has_space('buy'):
|
|
|
|
result_dict['params'].update({p.name: params.get(p.name)
|
2019-08-15 20:13:46 +00:00
|
|
|
for p in self.hyperopt_space('buy')})
|
2019-08-15 18:39:04 +00:00
|
|
|
if self.has_space('sell'):
|
|
|
|
result_dict['params'].update({p.name: params.get(p.name)
|
2019-08-15 20:13:46 +00:00
|
|
|
for p in self.hyperopt_space('sell')})
|
2019-08-15 18:39:04 +00:00
|
|
|
if self.has_space('roi'):
|
|
|
|
# Convert keys in min_roi dict to strings because
|
|
|
|
# rapidjson cannot dump dicts with integer keys...
|
|
|
|
# OrderedDict is used to keep the numeric order of the items
|
|
|
|
# in the dict.
|
2019-08-15 20:13:46 +00:00
|
|
|
result_dict['minimal_roi'] = OrderedDict(
|
|
|
|
(str(k), v) for k, v in self.custom_hyperopt.generate_roi_table(params).items()
|
|
|
|
)
|
2019-08-15 18:39:04 +00:00
|
|
|
if self.has_space('stoploss'):
|
|
|
|
result_dict['stoploss'] = params.get('stoploss')
|
|
|
|
print(rapidjson.dumps(result_dict, default=str, number_mode=rapidjson.NM_NATIVE))
|
|
|
|
else:
|
|
|
|
if self.has_space('buy'):
|
|
|
|
print('Buy hyperspace params:')
|
|
|
|
pprint({p.name: params.get(p.name) for p in self.hyperopt_space('buy')},
|
|
|
|
indent=4)
|
|
|
|
if self.has_space('sell'):
|
|
|
|
print('Sell hyperspace params:')
|
|
|
|
pprint({p.name: params.get(p.name) for p in self.hyperopt_space('sell')},
|
|
|
|
indent=4)
|
|
|
|
if self.has_space('roi'):
|
|
|
|
print("ROI table:")
|
2019-08-20 20:00:23 +00:00
|
|
|
# Round printed values to 5 digits after the decimal point
|
|
|
|
pprint(round_dict(self.custom_hyperopt.generate_roi_table(params), 5), indent=4)
|
2019-08-15 18:39:04 +00:00
|
|
|
if self.has_space('stoploss'):
|
2019-08-20 20:00:23 +00:00
|
|
|
# Also round to 5 digits after the decimal point
|
|
|
|
print(f"Stoploss: {round(params.get('stoploss'), 5)}")
|
2018-03-02 13:46:32 +00:00
|
|
|
|
2019-11-23 08:32:33 +00:00
|
|
|
def is_best(self, results) -> bool:
|
|
|
|
return results['loss'] < self.current_best_loss
|
|
|
|
|
2018-03-02 13:46:32 +00:00
|
|
|
def log_results(self, results) -> None:
|
|
|
|
"""
|
|
|
|
Log results if it is better than any previous evaluation
|
|
|
|
"""
|
2019-05-10 07:54:44 +00:00
|
|
|
print_all = self.config.get('print_all', False)
|
2019-11-23 08:32:33 +00:00
|
|
|
is_best_loss = self.is_best(results)
|
|
|
|
|
|
|
|
if not print_all:
|
2019-11-23 08:51:52 +00:00
|
|
|
print('.', end='' if results['current_epoch'] % 100 != 0 else None) # type: ignore
|
2019-11-23 08:32:33 +00:00
|
|
|
sys.stdout.flush()
|
|
|
|
|
2019-08-03 16:09:42 +00:00
|
|
|
if print_all or is_best_loss:
|
|
|
|
if is_best_loss:
|
|
|
|
self.current_best_loss = results['loss']
|
2019-07-30 08:47:28 +00:00
|
|
|
log_str = self.format_results_logstring(results)
|
2019-08-03 16:09:42 +00:00
|
|
|
# Colorize output
|
|
|
|
if self.config.get('print_colorized', False):
|
|
|
|
if results['total_profit'] > 0:
|
2019-08-09 11:48:57 +00:00
|
|
|
log_str = Fore.GREEN + log_str
|
2019-08-10 18:59:44 +00:00
|
|
|
if print_all and is_best_loss:
|
|
|
|
log_str = Style.BRIGHT + log_str
|
2019-05-10 07:54:44 +00:00
|
|
|
if print_all:
|
2019-07-30 08:47:28 +00:00
|
|
|
print(log_str)
|
2019-05-10 07:54:44 +00:00
|
|
|
else:
|
2019-11-06 18:33:15 +00:00
|
|
|
print(f'\n{log_str}')
|
2018-03-02 13:46:32 +00:00
|
|
|
|
2019-07-30 08:47:28 +00:00
|
|
|
def format_results_logstring(self, results) -> str:
|
2019-11-23 08:32:33 +00:00
|
|
|
current = results['current_epoch']
|
2019-07-30 08:47:28 +00:00
|
|
|
total = self.total_epochs
|
|
|
|
res = results['results_explanation']
|
|
|
|
loss = results['loss']
|
|
|
|
log_str = f'{current:5d}/{total}: {res} Objective: {loss:.5f}'
|
|
|
|
log_str = f'*{log_str}' if results['is_initial_point'] else f' {log_str}'
|
|
|
|
return log_str
|
|
|
|
|
2018-03-17 21:43:36 +00:00
|
|
|
def has_space(self, space: str) -> bool:
|
2018-03-04 08:51:22 +00:00
|
|
|
"""
|
|
|
|
Tell if a space value is contained in the configuration
|
|
|
|
"""
|
2019-08-01 20:57:26 +00:00
|
|
|
return any(s in self.config['spaces'] for s in [space, 'all'])
|
2018-03-04 08:51:22 +00:00
|
|
|
|
2019-08-02 19:22:58 +00:00
|
|
|
def hyperopt_space(self, space: Optional[str] = None) -> List[Dimension]:
|
2018-03-02 13:46:32 +00:00
|
|
|
"""
|
2019-08-03 07:20:20 +00:00
|
|
|
Return the dimensions in the hyperoptimization space.
|
|
|
|
:param space: Defines hyperspace to return dimensions for.
|
|
|
|
If None, then the self.has_space() will be used to return dimensions
|
2019-08-02 19:22:58 +00:00
|
|
|
for all hyperspaces used.
|
2018-03-02 13:46:32 +00:00
|
|
|
"""
|
2018-06-22 04:10:37 +00:00
|
|
|
spaces: List[Dimension] = []
|
2019-08-02 19:22:58 +00:00
|
|
|
if space == 'buy' or (space is None and self.has_space('buy')):
|
2019-08-01 20:57:26 +00:00
|
|
|
logger.debug("Hyperopt has 'buy' space")
|
2018-11-07 18:46:04 +00:00
|
|
|
spaces += self.custom_hyperopt.indicator_space()
|
2019-08-02 19:22:58 +00:00
|
|
|
if space == 'sell' or (space is None and self.has_space('sell')):
|
2019-08-01 20:57:26 +00:00
|
|
|
logger.debug("Hyperopt has 'sell' space")
|
2019-01-06 09:16:30 +00:00
|
|
|
spaces += self.custom_hyperopt.sell_indicator_space()
|
2019-08-02 19:22:58 +00:00
|
|
|
if space == 'roi' or (space is None and self.has_space('roi')):
|
2019-08-01 20:57:26 +00:00
|
|
|
logger.debug("Hyperopt has 'roi' space")
|
2018-11-07 18:46:04 +00:00
|
|
|
spaces += self.custom_hyperopt.roi_space()
|
2019-08-02 19:22:58 +00:00
|
|
|
if space == 'stoploss' or (space is None and self.has_space('stoploss')):
|
2019-08-01 20:57:26 +00:00
|
|
|
logger.debug("Hyperopt has 'stoploss' space")
|
2018-11-07 18:46:04 +00:00
|
|
|
spaces += self.custom_hyperopt.stoploss_space()
|
2018-06-22 04:10:37 +00:00
|
|
|
return spaces
|
2017-12-26 08:08:10 +00:00
|
|
|
|
2019-09-23 08:59:34 +00:00
|
|
|
def generate_optimizer(self, _params: Dict, iteration=None) -> Dict:
|
2019-07-15 18:28:55 +00:00
|
|
|
"""
|
|
|
|
Used Optimize function. Called once per epoch to optimize whatever is configured.
|
|
|
|
Keep this function as optimized as possible!
|
|
|
|
"""
|
2018-11-07 18:46:04 +00:00
|
|
|
params = self.get_args(_params)
|
2019-09-23 08:59:34 +00:00
|
|
|
|
2018-03-04 09:33:39 +00:00
|
|
|
if self.has_space('roi'):
|
2019-09-18 19:57:17 +00:00
|
|
|
self.backtesting.strategy.minimal_roi = \
|
|
|
|
self.custom_hyperopt.generate_roi_table(params)
|
2018-03-02 13:46:32 +00:00
|
|
|
|
2018-03-04 09:33:39 +00:00
|
|
|
if self.has_space('buy'):
|
2019-09-18 19:57:17 +00:00
|
|
|
self.backtesting.strategy.advise_buy = \
|
|
|
|
self.custom_hyperopt.buy_strategy_generator(params)
|
2018-03-04 08:51:22 +00:00
|
|
|
|
2019-01-06 09:16:30 +00:00
|
|
|
if self.has_space('sell'):
|
2019-09-18 19:57:17 +00:00
|
|
|
self.backtesting.strategy.advise_sell = \
|
|
|
|
self.custom_hyperopt.sell_strategy_generator(params)
|
2019-01-06 09:16:30 +00:00
|
|
|
|
2018-03-04 09:33:39 +00:00
|
|
|
if self.has_space('stoploss'):
|
2019-08-23 21:10:35 +00:00
|
|
|
self.backtesting.strategy.stoploss = params['stoploss']
|
2018-03-02 13:46:32 +00:00
|
|
|
|
2019-07-21 14:07:06 +00:00
|
|
|
processed = load(self.tickerdata_pickle)
|
2019-07-14 17:56:17 +00:00
|
|
|
|
2018-11-04 12:43:09 +00:00
|
|
|
min_date, max_date = get_timeframe(processed)
|
2019-07-14 17:56:17 +00:00
|
|
|
|
2019-08-23 21:10:35 +00:00
|
|
|
results = self.backtesting.backtest(
|
2018-03-02 13:46:32 +00:00
|
|
|
{
|
|
|
|
'stake_amount': self.config['stake_amount'],
|
2018-07-03 11:46:16 +00:00
|
|
|
'processed': processed,
|
2019-07-15 18:28:55 +00:00
|
|
|
'max_open_trades': self.max_open_trades,
|
2019-07-16 03:50:27 +00:00
|
|
|
'position_stacking': self.position_stacking,
|
2018-10-16 17:35:16 +00:00
|
|
|
'start_date': min_date,
|
|
|
|
'end_date': max_date,
|
2018-03-02 13:46:32 +00:00
|
|
|
}
|
|
|
|
)
|
2019-07-30 08:47:28 +00:00
|
|
|
results_explanation = self.format_results(results)
|
2018-03-02 13:46:32 +00:00
|
|
|
|
|
|
|
trade_count = len(results.index)
|
2019-08-03 16:09:42 +00:00
|
|
|
total_profit = results.profit_abs.sum()
|
2018-03-02 13:46:32 +00:00
|
|
|
|
2019-05-12 18:14:00 +00:00
|
|
|
# If this evaluation contains too short amount of trades to be
|
|
|
|
# interesting -- consider it as 'bad' (assigned max. loss value)
|
2019-05-01 12:27:58 +00:00
|
|
|
# in order to cast this hyperspace point away from optimization
|
|
|
|
# path. We do not want to optimize 'hodl' strategies.
|
|
|
|
if trade_count < self.config['hyperopt_min_trades']:
|
2018-07-02 08:44:33 +00:00
|
|
|
return {
|
|
|
|
'loss': MAX_LOSS,
|
|
|
|
'params': params,
|
2019-07-30 08:47:28 +00:00
|
|
|
'results_explanation': results_explanation,
|
2019-08-03 16:09:42 +00:00
|
|
|
'total_profit': total_profit,
|
2018-07-02 08:44:33 +00:00
|
|
|
}
|
|
|
|
|
2019-07-15 19:35:42 +00:00
|
|
|
loss = self.calculate_loss(results=results, trade_count=trade_count,
|
|
|
|
min_date=min_date.datetime, max_date=max_date.datetime)
|
2018-03-02 13:46:32 +00:00
|
|
|
|
2018-06-19 18:57:42 +00:00
|
|
|
return {
|
|
|
|
'loss': loss,
|
2018-06-22 10:02:26 +00:00
|
|
|
'params': params,
|
2019-07-30 08:47:28 +00:00
|
|
|
'results_explanation': results_explanation,
|
2019-08-03 16:09:42 +00:00
|
|
|
'total_profit': total_profit,
|
2018-06-19 18:57:42 +00:00
|
|
|
}
|
2017-11-25 00:04:11 +00:00
|
|
|
|
2018-06-02 21:07:31 +00:00
|
|
|
def format_results(self, results: DataFrame) -> str:
|
2018-03-02 13:46:32 +00:00
|
|
|
"""
|
2019-07-30 08:47:28 +00:00
|
|
|
Return the formatted results explanation in a string
|
2018-03-02 13:46:32 +00:00
|
|
|
"""
|
2018-06-14 05:52:13 +00:00
|
|
|
trades = len(results.index)
|
2019-05-12 18:14:00 +00:00
|
|
|
avg_profit = results.profit_percent.mean() * 100.0
|
2018-06-14 05:52:13 +00:00
|
|
|
total_profit = results.profit_abs.sum()
|
|
|
|
stake_cur = self.config['stake_currency']
|
2019-05-12 18:14:00 +00:00
|
|
|
profit = results.profit_percent.sum() * 100.0
|
2018-06-14 05:52:13 +00:00
|
|
|
duration = results.trade_duration.mean()
|
|
|
|
|
2019-05-12 18:14:00 +00:00
|
|
|
return (f'{trades:6d} trades. Avg profit {avg_profit: 5.2f}%. '
|
2018-06-14 05:52:13 +00:00
|
|
|
f'Total profit {total_profit: 11.8f} {stake_cur} '
|
2019-11-06 18:33:15 +00:00
|
|
|
f'({profit: 7.2f}\N{GREEK CAPITAL LETTER SIGMA}%). '
|
|
|
|
f'Avg duration {duration:5.1f} mins.'
|
|
|
|
).encode(locale.getpreferredencoding(), 'replace').decode('utf-8')
|
2018-03-02 13:46:32 +00:00
|
|
|
|
2019-09-16 18:22:07 +00:00
|
|
|
def get_optimizer(self, dimensions, cpu_count) -> Optimizer:
|
2018-06-24 12:27:53 +00:00
|
|
|
return Optimizer(
|
2019-09-16 18:22:07 +00:00
|
|
|
dimensions,
|
2018-06-24 12:27:53 +00:00
|
|
|
base_estimator="ET",
|
|
|
|
acq_optimizer="auto",
|
2019-05-10 07:54:44 +00:00
|
|
|
n_initial_points=INITIAL_POINTS,
|
2019-04-23 18:18:52 +00:00
|
|
|
acq_optimizer_kwargs={'n_jobs': cpu_count},
|
|
|
|
random_state=self.config.get('hyperopt_random_state', None)
|
2018-06-24 12:27:53 +00:00
|
|
|
)
|
|
|
|
|
2019-09-23 08:59:34 +00:00
|
|
|
def fix_optimizer_models_list(self):
|
|
|
|
"""
|
|
|
|
WORKAROUND: Since skopt is not actively supported, this resolves problems with skopt
|
|
|
|
memory usage, see also: https://github.com/scikit-optimize/scikit-optimize/pull/746
|
|
|
|
|
|
|
|
This may cease working when skopt updates if implementation of this intrinsic
|
|
|
|
part changes.
|
|
|
|
"""
|
|
|
|
n = len(self.opt.models) - SKOPT_MODELS_MAX_NUM
|
|
|
|
# Keep no more than 2*SKOPT_MODELS_MAX_NUM models in the skopt models list,
|
2019-09-23 10:25:31 +00:00
|
|
|
# remove the old ones. These are actually of no use, the current model
|
|
|
|
# from the estimator is the only one used in the skopt optimizer.
|
|
|
|
# Freqtrade code also does not inspect details of the models.
|
2019-09-23 08:59:34 +00:00
|
|
|
if n >= SKOPT_MODELS_MAX_NUM:
|
|
|
|
logger.debug(f"Fixing skopt models list, removing {n} old items...")
|
|
|
|
del self.opt.models[0:n]
|
|
|
|
|
|
|
|
def run_optimizer_parallel(self, parallel, asked, i) -> List:
|
2018-11-20 16:43:49 +00:00
|
|
|
return parallel(delayed(
|
2019-09-23 08:59:34 +00:00
|
|
|
wrap_non_picklable_objects(self.generate_optimizer))(v, i) for v in asked)
|
2018-06-24 12:27:53 +00:00
|
|
|
|
2018-06-25 08:38:14 +00:00
|
|
|
def load_previous_results(self):
|
|
|
|
""" read trials file if we have one """
|
2019-07-21 13:56:44 +00:00
|
|
|
if self.trials_file.is_file() and self.trials_file.stat().st_size > 0:
|
2018-06-25 08:38:14 +00:00
|
|
|
self.trials = self.read_trials()
|
|
|
|
logger.info(
|
|
|
|
'Loaded %d previous evaluations from disk.',
|
|
|
|
len(self.trials)
|
|
|
|
)
|
|
|
|
|
2018-03-17 21:43:36 +00:00
|
|
|
def start(self) -> None:
|
2019-10-23 18:13:43 +00:00
|
|
|
data, timerange = self.backtesting.load_bt_data()
|
2017-11-25 00:04:11 +00:00
|
|
|
|
2019-10-23 18:13:43 +00:00
|
|
|
preprocessed = self.backtesting.strategy.tickerdata_to_dataframe(data)
|
2019-05-13 20:56:59 +00:00
|
|
|
|
2019-10-23 18:13:43 +00:00
|
|
|
# Trim startup period from analyzed dataframe
|
|
|
|
for pair, df in preprocessed.items():
|
|
|
|
preprocessed[pair] = trim_dataframe(df, timerange)
|
2019-05-13 20:56:59 +00:00
|
|
|
min_date, max_date = get_timeframe(data)
|
2019-06-15 11:46:19 +00:00
|
|
|
|
2019-05-13 20:56:59 +00:00
|
|
|
logger.info(
|
2019-05-15 09:05:35 +00:00
|
|
|
'Hyperopting with data from %s up to %s (%s days)..',
|
2019-10-23 18:13:43 +00:00
|
|
|
min_date.isoformat(), max_date.isoformat(), (max_date - min_date).days
|
2019-05-13 20:56:59 +00:00
|
|
|
)
|
2019-07-21 14:07:06 +00:00
|
|
|
dump(preprocessed, self.tickerdata_pickle)
|
2019-04-22 18:24:45 +00:00
|
|
|
|
|
|
|
# We don't need exchange instance anymore while running hyperopt
|
2019-08-23 21:10:35 +00:00
|
|
|
self.backtesting.exchange = None # type: ignore
|
2019-04-22 18:24:45 +00:00
|
|
|
|
2018-06-25 08:38:14 +00:00
|
|
|
self.load_previous_results()
|
2018-03-02 13:46:32 +00:00
|
|
|
|
2019-04-23 18:25:36 +00:00
|
|
|
cpus = cpu_count()
|
2019-08-25 18:38:51 +00:00
|
|
|
logger.info(f"Found {cpus} CPU cores. Let's make them scream!")
|
2019-04-22 21:30:09 +00:00
|
|
|
config_jobs = self.config.get('hyperopt_jobs', -1)
|
|
|
|
logger.info(f'Number of parallel jobs set as: {config_jobs}')
|
2018-06-21 11:59:36 +00:00
|
|
|
|
2019-09-16 18:22:07 +00:00
|
|
|
self.dimensions = self.hyperopt_space()
|
|
|
|
self.opt = self.get_optimizer(self.dimensions, config_jobs)
|
2019-08-04 19:54:19 +00:00
|
|
|
|
2019-08-09 11:48:57 +00:00
|
|
|
if self.config.get('print_colorized', False):
|
|
|
|
colorama_init(autoreset=True)
|
2019-08-04 19:54:19 +00:00
|
|
|
|
2018-06-22 10:02:26 +00:00
|
|
|
try:
|
2019-04-22 21:30:09 +00:00
|
|
|
with Parallel(n_jobs=config_jobs) as parallel:
|
|
|
|
jobs = parallel._effective_n_jobs()
|
|
|
|
logger.info(f'Effective number of parallel workers used: {jobs}')
|
2019-07-30 08:47:28 +00:00
|
|
|
EVALS = max(self.total_epochs // jobs, 1)
|
2018-07-03 18:54:32 +00:00
|
|
|
for i in range(EVALS):
|
2019-09-16 18:22:07 +00:00
|
|
|
asked = self.opt.ask(n_points=jobs)
|
2019-09-23 08:59:34 +00:00
|
|
|
f_val = self.run_optimizer_parallel(parallel, asked, i)
|
2019-09-16 18:22:07 +00:00
|
|
|
self.opt.tell(asked, [v['loss'] for v in f_val])
|
2019-09-23 08:59:34 +00:00
|
|
|
self.fix_optimizer_models_list()
|
2019-04-22 21:30:09 +00:00
|
|
|
for j in range(jobs):
|
2019-11-23 08:32:33 +00:00
|
|
|
# Use human-friendly index here (starting from 1)
|
|
|
|
current = i * jobs + j + 1
|
2019-07-30 08:47:28 +00:00
|
|
|
val = f_val[j]
|
|
|
|
val['current_epoch'] = current
|
2019-11-23 08:32:33 +00:00
|
|
|
val['is_initial_point'] = current <= INITIAL_POINTS
|
|
|
|
logger.debug(f"Optimizer epoch evaluated: {val}")
|
|
|
|
is_best = self.is_best(val)
|
2019-07-30 08:47:28 +00:00
|
|
|
self.log_results(val)
|
|
|
|
self.trials.append(val)
|
2019-11-23 08:32:33 +00:00
|
|
|
if is_best or current % 100 == 0:
|
|
|
|
self.save_trials()
|
2018-06-22 10:02:26 +00:00
|
|
|
except KeyboardInterrupt:
|
|
|
|
print('User interrupted..')
|
2018-01-07 01:12:32 +00:00
|
|
|
|
2019-11-23 08:32:33 +00:00
|
|
|
self.save_trials(final=True)
|
2018-03-02 13:46:32 +00:00
|
|
|
self.log_trials_result()
|