2022-05-03 08:14:17 +00:00
|
|
|
import copy
|
2022-05-04 15:53:40 +00:00
|
|
|
import logging
|
2022-05-05 13:35:51 +00:00
|
|
|
import shutil
|
2022-09-03 13:52:29 +00:00
|
|
|
from datetime import datetime, timezone
|
2022-09-06 17:46:58 +00:00
|
|
|
from math import cos, sin
|
2022-05-04 15:42:34 +00:00
|
|
|
from pathlib import Path
|
2022-08-10 15:43:06 +00:00
|
|
|
from typing import Any, Dict, List, Tuple
|
2022-05-04 15:42:34 +00:00
|
|
|
|
2022-05-03 08:14:17 +00:00
|
|
|
import numpy as np
|
2022-05-06 14:20:52 +00:00
|
|
|
import numpy.typing as npt
|
2022-05-03 08:14:17 +00:00
|
|
|
import pandas as pd
|
|
|
|
from pandas import DataFrame
|
2022-09-06 18:10:12 +00:00
|
|
|
from scipy import stats
|
2022-05-22 15:51:49 +00:00
|
|
|
from sklearn import linear_model
|
2022-08-04 15:00:59 +00:00
|
|
|
from sklearn.cluster import DBSCAN
|
2022-05-03 08:14:17 +00:00
|
|
|
from sklearn.metrics.pairwise import pairwise_distances
|
2022-05-04 15:42:34 +00:00
|
|
|
from sklearn.model_selection import train_test_split
|
2022-08-06 05:25:59 +00:00
|
|
|
from sklearn.neighbors import NearestNeighbors
|
2022-05-04 15:42:34 +00:00
|
|
|
|
2022-05-03 08:14:17 +00:00
|
|
|
from freqtrade.configuration import TimeRange
|
2022-09-18 11:20:36 +00:00
|
|
|
from freqtrade.constants import Config
|
2022-05-26 22:43:52 +00:00
|
|
|
from freqtrade.exceptions import OperationalException
|
2022-08-09 13:30:25 +00:00
|
|
|
from freqtrade.exchange import timeframe_to_seconds
|
2022-09-26 02:14:00 +00:00
|
|
|
from freqtrade.freqai import freqai_util
|
2022-05-09 13:25:00 +00:00
|
|
|
from freqtrade.strategy.interface import IStrategy
|
2022-08-04 15:00:59 +00:00
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
|
2022-05-03 08:14:17 +00:00
|
|
|
SECONDS_IN_DAY = 86400
|
2022-07-10 10:34:09 +00:00
|
|
|
SECONDS_IN_HOUR = 3600
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-05-04 15:53:40 +00:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
|
2022-05-06 10:54:49 +00:00
|
|
|
class FreqaiDataKitchen:
|
2022-05-03 08:14:17 +00:00
|
|
|
"""
|
2022-05-23 19:05:05 +00:00
|
|
|
Class designed to analyze data for a single pair. Employed by the IFreqaiModel class.
|
2022-05-03 08:14:17 +00:00
|
|
|
Functionalities include holding, saving, loading, and analyzing the data.
|
2022-07-23 11:32:04 +00:00
|
|
|
|
2022-08-14 18:24:29 +00:00
|
|
|
This object is not persistent, it is reinstantiated for each coin, each time the coin
|
|
|
|
model needs to be inferenced or trained.
|
|
|
|
|
2022-07-23 11:32:04 +00:00
|
|
|
Record of contribution:
|
|
|
|
FreqAI was developed by a group of individuals who all contributed specific skillsets to the
|
|
|
|
project.
|
|
|
|
|
|
|
|
Conception and software development:
|
|
|
|
Robert Caulk @robcaulk
|
|
|
|
|
|
|
|
Theoretical brainstorming:
|
2022-08-02 18:14:02 +00:00
|
|
|
Elin Törnquist @th0rntwig
|
2022-07-23 11:32:04 +00:00
|
|
|
|
|
|
|
Code review, software architecture brainstorming:
|
|
|
|
@xmatthias
|
|
|
|
|
|
|
|
Beta testing and bug reporting:
|
|
|
|
@bloodhunter4rc, Salah Lamkadem @ikonx, @ken11o2, @longyu, @paranoidandy, @smidelis, @smarm
|
2022-08-14 18:24:29 +00:00
|
|
|
Juha Nykänen @suikula, Wagner Costa @wagnercosta, Johan Vlugt @Jooopieeert
|
2022-05-03 08:14:17 +00:00
|
|
|
"""
|
|
|
|
|
2022-07-03 08:59:38 +00:00
|
|
|
def __init__(
|
|
|
|
self,
|
2022-09-18 11:20:36 +00:00
|
|
|
config: Config,
|
2022-07-03 08:59:38 +00:00
|
|
|
live: bool = False,
|
|
|
|
pair: str = "",
|
|
|
|
):
|
2022-08-09 13:30:25 +00:00
|
|
|
self.data: Dict[str, Any] = {}
|
|
|
|
self.data_dictionary: Dict[str, DataFrame] = {}
|
2022-05-03 08:14:17 +00:00
|
|
|
self.config = config
|
2022-08-06 12:55:46 +00:00
|
|
|
self.freqai_config: Dict[str, Any] = config["freqai"]
|
2022-07-03 15:34:44 +00:00
|
|
|
self.full_df: DataFrame = DataFrame()
|
|
|
|
self.append_df: DataFrame = DataFrame()
|
2022-05-23 19:05:05 +00:00
|
|
|
self.data_path = Path()
|
2022-07-02 16:09:38 +00:00
|
|
|
self.label_list: List = []
|
2022-07-26 08:24:14 +00:00
|
|
|
self.training_features_list: List = []
|
2022-05-09 13:25:00 +00:00
|
|
|
self.model_filename: str = ""
|
2022-09-01 10:09:23 +00:00
|
|
|
self.backtesting_results_path = Path()
|
2022-09-03 12:00:01 +00:00
|
|
|
self.backtest_predictions_folder: str = "backtesting_predictions"
|
2022-05-19 17:27:38 +00:00
|
|
|
self.live = live
|
2022-05-23 19:05:05 +00:00
|
|
|
self.pair = pair
|
2022-08-14 15:23:14 +00:00
|
|
|
|
2022-05-22 15:51:49 +00:00
|
|
|
self.svm_model: linear_model.SGDOneClassSVM = None
|
2022-08-06 12:55:46 +00:00
|
|
|
self.keras: bool = self.freqai_config.get("keras", False)
|
2022-06-03 13:19:46 +00:00
|
|
|
self.set_all_pairs()
|
2022-09-26 02:14:00 +00:00
|
|
|
self.backtest_live_models = config.get("freqai_backtest_live_models", False)
|
|
|
|
|
2022-05-19 17:27:38 +00:00
|
|
|
if not self.live:
|
2022-09-29 04:48:38 +00:00
|
|
|
self.full_path = freqai_util.get_full_models_path(self.config)
|
2022-07-03 08:59:38 +00:00
|
|
|
self.full_timerange = self.create_fulltimerange(
|
2022-08-06 12:55:46 +00:00
|
|
|
self.config["timerange"], self.freqai_config.get("train_period_days", 0)
|
2022-07-03 08:59:38 +00:00
|
|
|
)
|
2022-05-05 13:35:51 +00:00
|
|
|
|
2022-09-25 13:35:55 +00:00
|
|
|
if self.backtest_live_models:
|
2022-09-26 02:14:00 +00:00
|
|
|
self.set_timerange_from_ready_models()
|
2022-09-25 13:35:55 +00:00
|
|
|
(self.training_timeranges,
|
|
|
|
self.backtesting_timeranges) = self.split_timerange_live_models()
|
|
|
|
else:
|
|
|
|
(self.training_timeranges, self.backtesting_timeranges) = self.split_timerange(
|
|
|
|
self.full_timerange,
|
|
|
|
config["freqai"]["train_period_days"],
|
|
|
|
config["freqai"]["backtest_period_days"],
|
|
|
|
)
|
2022-07-21 20:11:46 +00:00
|
|
|
|
2022-08-02 18:14:02 +00:00
|
|
|
self.data['extra_returns_per_train'] = self.freqai_config.get('extra_returns_per_train', {})
|
2022-08-04 10:14:56 +00:00
|
|
|
self.thread_count = self.freqai_config.get("data_kitchen_thread_count", -1)
|
2022-08-09 13:30:25 +00:00
|
|
|
self.train_dates: DataFrame = pd.DataFrame()
|
2022-08-09 15:31:38 +00:00
|
|
|
self.unique_classes: Dict[str, list] = {}
|
2022-08-10 13:16:50 +00:00
|
|
|
self.unique_class_list: list = []
|
2022-09-24 16:01:53 +00:00
|
|
|
self.backtest_live_models_data: Dict[str, Any] = {}
|
2022-08-02 18:14:02 +00:00
|
|
|
|
2022-07-03 08:59:38 +00:00
|
|
|
def set_paths(
|
|
|
|
self,
|
|
|
|
pair: str,
|
|
|
|
trained_timestamp: int = None,
|
|
|
|
) -> None:
|
2022-06-03 13:19:46 +00:00
|
|
|
"""
|
|
|
|
Set the paths to the data for the present coin/botloop
|
2022-10-10 12:15:30 +00:00
|
|
|
:param metadata: dict = strategy furnished pair metadata
|
|
|
|
:param trained_timestamp: int = timestamp of most recent training
|
2022-06-03 13:19:46 +00:00
|
|
|
"""
|
2022-09-29 04:48:38 +00:00
|
|
|
self.full_path = freqai_util.get_full_models_path(self.config)
|
2022-07-03 08:59:38 +00:00
|
|
|
self.data_path = Path(
|
2022-07-21 09:25:28 +00:00
|
|
|
self.full_path
|
2022-08-02 18:14:02 +00:00
|
|
|
/ f"sub-train-{pair.split('/')[0]}_{trained_timestamp}"
|
2022-07-03 08:59:38 +00:00
|
|
|
)
|
2022-05-09 13:25:00 +00:00
|
|
|
|
|
|
|
return
|
2022-05-05 13:35:51 +00:00
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
def make_train_test_datasets(
|
|
|
|
self, filtered_dataframe: DataFrame, labels: DataFrame
|
|
|
|
) -> Dict[Any, Any]:
|
|
|
|
"""
|
|
|
|
Given the dataframe for the full history for training, split the data into
|
|
|
|
training and test data according to user specified parameters in configuration
|
|
|
|
file.
|
2022-10-10 12:15:30 +00:00
|
|
|
:param filtered_dataframe: cleaned dataframe ready to be split.
|
|
|
|
:param labels: cleaned labels ready to be split.
|
2022-05-04 15:42:34 +00:00
|
|
|
"""
|
2022-07-29 06:12:50 +00:00
|
|
|
feat_dict = self.freqai_config["feature_parameters"]
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-09-29 22:33:08 +00:00
|
|
|
if 'shuffle' not in self.freqai_config['data_split_parameters']:
|
|
|
|
self.freqai_config["data_split_parameters"].update({'shuffle': False})
|
2022-09-28 16:23:56 +00:00
|
|
|
|
2022-05-06 14:20:52 +00:00
|
|
|
weights: npt.ArrayLike
|
2022-07-10 10:34:09 +00:00
|
|
|
if feat_dict.get("weight_factor", 0) > 0:
|
2022-05-03 08:14:17 +00:00
|
|
|
weights = self.set_weights_higher_recent(len(filtered_dataframe))
|
2022-05-04 15:42:34 +00:00
|
|
|
else:
|
|
|
|
weights = np.ones(len(filtered_dataframe))
|
|
|
|
|
2022-07-25 17:40:13 +00:00
|
|
|
if self.freqai_config.get('data_split_parameters', {}).get('test_size', 0.1) != 0:
|
|
|
|
(
|
|
|
|
train_features,
|
|
|
|
test_features,
|
|
|
|
train_labels,
|
|
|
|
test_labels,
|
|
|
|
train_weights,
|
|
|
|
test_weights,
|
|
|
|
) = train_test_split(
|
|
|
|
filtered_dataframe[: filtered_dataframe.shape[0]],
|
|
|
|
labels,
|
|
|
|
weights,
|
|
|
|
**self.config["freqai"]["data_split_parameters"],
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
test_labels = np.zeros(2)
|
|
|
|
test_features = pd.DataFrame()
|
|
|
|
test_weights = np.zeros(2)
|
|
|
|
train_features = filtered_dataframe
|
|
|
|
train_labels = labels
|
|
|
|
train_weights = weights
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-08-29 09:04:58 +00:00
|
|
|
# Simplest way to reverse the order of training and test data:
|
|
|
|
if self.freqai_config['feature_parameters'].get('reverse_train_test_order', False):
|
|
|
|
return self.build_data_dictionary(
|
|
|
|
test_features, train_features, test_labels,
|
|
|
|
train_labels, test_weights, train_weights
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
return self.build_data_dictionary(
|
|
|
|
train_features, test_features, train_labels,
|
|
|
|
test_labels, train_weights, test_weights
|
|
|
|
)
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
def filter_features(
|
|
|
|
self,
|
2022-09-07 15:47:27 +00:00
|
|
|
unfiltered_df: DataFrame,
|
2022-05-04 15:42:34 +00:00
|
|
|
training_feature_list: List,
|
2022-07-02 16:09:38 +00:00
|
|
|
label_list: List = list(),
|
2022-05-04 15:42:34 +00:00
|
|
|
training_filter: bool = True,
|
|
|
|
) -> Tuple[DataFrame, DataFrame]:
|
|
|
|
"""
|
2022-07-02 16:09:38 +00:00
|
|
|
Filter the unfiltered dataframe to extract the user requested features/labels and properly
|
2022-05-04 15:42:34 +00:00
|
|
|
remove all NaNs. Any row with a NaN is removed from training dataset or replaced with
|
|
|
|
0s in the prediction dataset. However, prediction dataset do_predict will reflect any
|
2022-05-03 08:14:17 +00:00
|
|
|
row that had a NaN and will shield user from that prediction.
|
2022-10-10 12:15:30 +00:00
|
|
|
|
|
|
|
:param unfiltered_df: the full dataframe for the present training period
|
|
|
|
:param training_feature_list: list, the training feature list constructed by
|
|
|
|
self.build_feature_list() according to user specified
|
|
|
|
parameters in the configuration file.
|
|
|
|
:param labels: the labels for the dataset
|
|
|
|
:param training_filter: boolean which lets the function know if it is training data or
|
|
|
|
prediction data to be filtered.
|
2022-05-03 08:14:17 +00:00
|
|
|
:returns:
|
2022-09-07 15:47:27 +00:00
|
|
|
:filtered_df: dataframe cleaned of NaNs and only containing the user
|
2022-05-03 08:14:17 +00:00
|
|
|
requested feature set.
|
|
|
|
:labels: labels cleaned of NaNs.
|
2022-05-04 15:42:34 +00:00
|
|
|
"""
|
2022-09-07 15:47:27 +00:00
|
|
|
filtered_df = unfiltered_df.filter(training_feature_list, axis=1)
|
|
|
|
filtered_df = filtered_df.replace([np.inf, -np.inf], np.nan)
|
2022-07-02 16:09:38 +00:00
|
|
|
|
2022-09-30 13:43:05 +00:00
|
|
|
drop_index = pd.isnull(filtered_df).any(axis=1) # get the rows that have NaNs,
|
2022-05-04 15:42:34 +00:00
|
|
|
drop_index = drop_index.replace(True, 1).replace(False, 0) # pep8 requirement.
|
2022-08-06 12:55:46 +00:00
|
|
|
if (training_filter):
|
2022-09-09 18:37:21 +00:00
|
|
|
const_cols = list((filtered_df.nunique() == 1).loc[lambda x: x].index)
|
|
|
|
if const_cols:
|
|
|
|
filtered_df = filtered_df.filter(filtered_df.columns.difference(const_cols))
|
|
|
|
logger.warning(f"Removed features {const_cols} with constant values.")
|
2022-08-06 12:55:46 +00:00
|
|
|
# we don't care about total row number (total no. datapoints) in training, we only care
|
2022-05-04 15:42:34 +00:00
|
|
|
# about removing any row with NaNs
|
2022-08-09 13:30:25 +00:00
|
|
|
# if labels has multiple columns (user wants to train multiple modelEs), we detect here
|
2022-09-07 15:47:27 +00:00
|
|
|
labels = unfiltered_df.filter(label_list, axis=1)
|
2022-09-30 13:43:05 +00:00
|
|
|
drop_index_labels = pd.isnull(labels).any(axis=1)
|
2022-05-04 15:42:34 +00:00
|
|
|
drop_index_labels = drop_index_labels.replace(True, 1).replace(False, 0)
|
2022-09-07 15:47:27 +00:00
|
|
|
dates = unfiltered_df['date']
|
|
|
|
filtered_df = filtered_df[
|
2022-05-04 15:42:34 +00:00
|
|
|
(drop_index == 0) & (drop_index_labels == 0)
|
|
|
|
] # dropping values
|
|
|
|
labels = labels[
|
|
|
|
(drop_index == 0) & (drop_index_labels == 0)
|
|
|
|
] # assuming the labels depend entirely on the dataframe here.
|
2022-08-09 13:30:25 +00:00
|
|
|
self.train_dates = dates[
|
|
|
|
(drop_index == 0) & (drop_index_labels == 0)
|
|
|
|
]
|
2022-05-31 16:42:27 +00:00
|
|
|
logger.info(
|
2022-09-07 15:47:27 +00:00
|
|
|
f"dropped {len(unfiltered_df) - len(filtered_df)} training points"
|
|
|
|
f" due to NaNs in populated dataset {len(unfiltered_df)}."
|
2022-05-31 16:42:27 +00:00
|
|
|
)
|
2022-09-07 15:47:27 +00:00
|
|
|
if (1 - len(filtered_df) / len(unfiltered_df)) > 0.1 and self.live:
|
|
|
|
worst_indicator = str(unfiltered_df.count().idxmin())
|
2022-05-31 16:42:27 +00:00
|
|
|
logger.warning(
|
2022-09-07 15:47:27 +00:00
|
|
|
f" {(1 - len(filtered_df)/len(unfiltered_df)) * 100:.0f} percent "
|
2022-07-05 10:42:32 +00:00
|
|
|
" of training data dropped due to NaNs, model may perform inconsistent "
|
|
|
|
f"with expectations. Verify {worst_indicator}"
|
2022-05-31 16:42:27 +00:00
|
|
|
)
|
2022-05-04 15:42:34 +00:00
|
|
|
self.data["filter_drop_index_training"] = drop_index
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
else:
|
|
|
|
# we are backtesting so we need to preserve row number to send back to strategy,
|
|
|
|
# so now we use do_predict to avoid any prediction based on a NaN
|
2022-09-30 13:43:05 +00:00
|
|
|
drop_index = pd.isnull(filtered_df).any(axis=1)
|
2022-05-04 15:42:34 +00:00
|
|
|
self.data["filter_drop_index_prediction"] = drop_index
|
2022-09-07 15:47:27 +00:00
|
|
|
filtered_df.fillna(0, inplace=True)
|
2022-05-04 15:42:34 +00:00
|
|
|
# replacing all NaNs with zeros to avoid issues in 'prediction', but any prediction
|
|
|
|
# that was based on a single NaN is ultimately protected from buys with do_predict
|
2022-05-03 08:14:17 +00:00
|
|
|
drop_index = ~drop_index
|
2022-05-04 15:42:34 +00:00
|
|
|
self.do_predict = np.array(drop_index.replace(True, 1).replace(False, 0))
|
2022-06-26 17:02:17 +00:00
|
|
|
if (len(self.do_predict) - self.do_predict.sum()) > 0:
|
|
|
|
logger.info(
|
|
|
|
"dropped %s of %s prediction data points due to NaNs.",
|
|
|
|
len(self.do_predict) - self.do_predict.sum(),
|
2022-09-07 15:47:27 +00:00
|
|
|
len(filtered_df),
|
2022-06-26 17:02:17 +00:00
|
|
|
)
|
2022-07-02 16:09:38 +00:00
|
|
|
labels = []
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-09-07 15:47:27 +00:00
|
|
|
return filtered_df, labels
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
def build_data_dictionary(
|
|
|
|
self,
|
|
|
|
train_df: DataFrame,
|
|
|
|
test_df: DataFrame,
|
|
|
|
train_labels: DataFrame,
|
|
|
|
test_labels: DataFrame,
|
|
|
|
train_weights: Any,
|
|
|
|
test_weights: Any,
|
|
|
|
) -> Dict:
|
|
|
|
|
|
|
|
self.data_dictionary = {
|
|
|
|
"train_features": train_df,
|
|
|
|
"test_features": test_df,
|
|
|
|
"train_labels": train_labels,
|
|
|
|
"test_labels": test_labels,
|
|
|
|
"train_weights": train_weights,
|
|
|
|
"test_weights": test_weights,
|
2022-08-09 13:30:25 +00:00
|
|
|
"train_dates": self.train_dates
|
2022-05-04 15:42:34 +00:00
|
|
|
}
|
2022-05-03 08:14:17 +00:00
|
|
|
|
|
|
|
return self.data_dictionary
|
|
|
|
|
2022-07-23 15:14:11 +00:00
|
|
|
def normalize_data(self, data_dictionary: Dict) -> Dict[Any, Any]:
|
2022-05-04 15:42:34 +00:00
|
|
|
"""
|
2022-05-22 15:51:49 +00:00
|
|
|
Normalize all data in the data_dictionary according to the training dataset
|
2022-10-10 12:15:30 +00:00
|
|
|
:param data_dictionary: dictionary containing the cleaned and
|
|
|
|
split training/test data/labels
|
2022-05-03 08:14:17 +00:00
|
|
|
:returns:
|
|
|
|
:data_dictionary: updated dictionary with standardized values.
|
2022-05-04 15:42:34 +00:00
|
|
|
"""
|
2022-09-01 19:51:33 +00:00
|
|
|
|
2022-05-03 08:14:17 +00:00
|
|
|
# standardize the data by training stats
|
2022-05-22 15:51:49 +00:00
|
|
|
train_max = data_dictionary["train_features"].max()
|
|
|
|
train_min = data_dictionary["train_features"].min()
|
2022-07-03 08:59:38 +00:00
|
|
|
data_dictionary["train_features"] = (
|
|
|
|
2 * (data_dictionary["train_features"] - train_min) / (train_max - train_min) - 1
|
|
|
|
)
|
|
|
|
data_dictionary["test_features"] = (
|
|
|
|
2 * (data_dictionary["test_features"] - train_min) / (train_max - train_min) - 1
|
|
|
|
)
|
2022-05-22 15:51:49 +00:00
|
|
|
|
|
|
|
for item in train_max.keys():
|
|
|
|
self.data[item + "_max"] = train_max[item]
|
|
|
|
self.data[item + "_min"] = train_min[item]
|
|
|
|
|
2022-07-23 15:14:11 +00:00
|
|
|
for item in data_dictionary["train_labels"].keys():
|
2022-07-09 08:13:33 +00:00
|
|
|
if data_dictionary["train_labels"][item].dtype == object:
|
2022-07-23 15:14:11 +00:00
|
|
|
continue
|
|
|
|
train_labels_max = data_dictionary["train_labels"][item].max()
|
|
|
|
train_labels_min = data_dictionary["train_labels"][item].min()
|
2022-07-24 06:42:50 +00:00
|
|
|
data_dictionary["train_labels"][item] = (
|
2022-07-23 14:14:13 +00:00
|
|
|
2
|
2022-07-23 15:14:11 +00:00
|
|
|
* (data_dictionary["train_labels"][item] - train_labels_min)
|
2022-07-23 14:14:13 +00:00
|
|
|
/ (train_labels_max - train_labels_min)
|
|
|
|
- 1
|
|
|
|
)
|
2022-07-25 17:40:13 +00:00
|
|
|
if self.freqai_config.get('data_split_parameters', {}).get('test_size', 0.1) != 0:
|
|
|
|
data_dictionary["test_labels"][item] = (
|
|
|
|
2
|
|
|
|
* (data_dictionary["test_labels"][item] - train_labels_min)
|
|
|
|
/ (train_labels_max - train_labels_min)
|
|
|
|
- 1
|
|
|
|
)
|
2022-07-23 14:14:13 +00:00
|
|
|
|
2022-09-03 17:48:30 +00:00
|
|
|
self.data[f"{item}_max"] = train_labels_max
|
|
|
|
self.data[f"{item}_min"] = train_labels_min
|
2022-05-22 15:51:49 +00:00
|
|
|
return data_dictionary
|
|
|
|
|
2022-09-03 14:54:30 +00:00
|
|
|
def normalize_single_dataframe(self, df: DataFrame) -> DataFrame:
|
|
|
|
|
|
|
|
train_max = df.max()
|
|
|
|
train_min = df.min()
|
|
|
|
df = (
|
|
|
|
2 * (df - train_min) / (train_max - train_min) - 1
|
|
|
|
)
|
|
|
|
|
|
|
|
for item in train_max.keys():
|
|
|
|
self.data[item + "_max"] = train_max[item]
|
|
|
|
self.data[item + "_min"] = train_min[item]
|
|
|
|
|
|
|
|
return df
|
|
|
|
|
2022-05-26 19:07:50 +00:00
|
|
|
def normalize_data_from_metadata(self, df: DataFrame) -> DataFrame:
|
2022-05-04 15:42:34 +00:00
|
|
|
"""
|
2022-05-26 19:07:50 +00:00
|
|
|
Normalize a set of data using the mean and standard deviation from
|
2022-05-03 08:14:17 +00:00
|
|
|
the associated training data.
|
2022-08-06 12:55:46 +00:00
|
|
|
:param df: Dataframe to be standardized
|
2022-05-04 15:42:34 +00:00
|
|
|
"""
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-05-22 15:51:49 +00:00
|
|
|
for item in df.keys():
|
2022-07-03 08:59:38 +00:00
|
|
|
df[item] = (
|
|
|
|
2
|
2022-07-24 06:42:50 +00:00
|
|
|
* (df[item] - self.data[f"{item}_min"])
|
|
|
|
/ (self.data[f"{item}_max"] - self.data[f"{item}_min"])
|
2022-07-03 08:59:38 +00:00
|
|
|
- 1
|
|
|
|
)
|
2022-05-22 15:51:49 +00:00
|
|
|
|
|
|
|
return df
|
|
|
|
|
2022-07-23 15:14:11 +00:00
|
|
|
def denormalize_labels_from_metadata(self, df: DataFrame) -> DataFrame:
|
|
|
|
"""
|
2022-09-03 13:52:29 +00:00
|
|
|
Denormalize a set of data using the mean and standard deviation from
|
2022-07-23 15:14:11 +00:00
|
|
|
the associated training data.
|
2022-08-06 12:55:46 +00:00
|
|
|
:param df: Dataframe of predictions to be denormalized
|
2022-07-23 15:14:11 +00:00
|
|
|
"""
|
|
|
|
|
2022-08-06 11:51:19 +00:00
|
|
|
for label in df.columns:
|
2022-08-10 13:16:50 +00:00
|
|
|
if df[label].dtype == object or label in self.unique_class_list:
|
2022-07-23 15:14:11 +00:00
|
|
|
continue
|
|
|
|
df[label] = (
|
|
|
|
(df[label] + 1)
|
|
|
|
* (self.data[f"{label}_max"] - self.data[f"{label}_min"])
|
|
|
|
/ 2
|
|
|
|
) + self.data[f"{label}_min"]
|
|
|
|
|
|
|
|
return df
|
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
def split_timerange(
|
2022-08-09 13:30:25 +00:00
|
|
|
self, tr: str, train_split: int = 28, bt_split: float = 7
|
2022-05-04 15:42:34 +00:00
|
|
|
) -> Tuple[list, list]:
|
|
|
|
"""
|
2022-05-03 08:14:17 +00:00
|
|
|
Function which takes a single time range (tr) and splits it
|
|
|
|
into sub timeranges to train and backtest on based on user input
|
|
|
|
tr: str, full timerange to train on
|
|
|
|
train_split: the period length for the each training (days). Specified in user
|
|
|
|
configuration file
|
2022-08-09 13:30:25 +00:00
|
|
|
bt_split: the backtesting length (days). Specified in user configuration file
|
2022-05-04 15:42:34 +00:00
|
|
|
"""
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-07-19 14:16:44 +00:00
|
|
|
if not isinstance(train_split, int) or train_split < 1:
|
|
|
|
raise OperationalException(
|
2022-08-06 12:55:46 +00:00
|
|
|
f"train_period_days must be an integer greater than 0. Got {train_split}."
|
2022-07-19 14:16:44 +00:00
|
|
|
)
|
2022-07-10 10:34:09 +00:00
|
|
|
train_period_days = train_split * SECONDS_IN_DAY
|
2022-05-03 08:14:17 +00:00
|
|
|
bt_period = bt_split * SECONDS_IN_DAY
|
|
|
|
|
|
|
|
full_timerange = TimeRange.parse_timerange(tr)
|
2022-05-06 10:54:49 +00:00
|
|
|
config_timerange = TimeRange.parse_timerange(self.config["timerange"])
|
2022-05-17 17:50:06 +00:00
|
|
|
if config_timerange.stopts == 0:
|
2022-07-03 08:59:38 +00:00
|
|
|
config_timerange.stopts = int(
|
2022-09-03 13:52:29 +00:00
|
|
|
datetime.now(tz=timezone.utc).timestamp()
|
2022-07-03 08:59:38 +00:00
|
|
|
)
|
2022-05-03 08:14:17 +00:00
|
|
|
timerange_train = copy.deepcopy(full_timerange)
|
|
|
|
timerange_backtest = copy.deepcopy(full_timerange)
|
|
|
|
|
|
|
|
tr_training_list = []
|
|
|
|
tr_backtesting_list = []
|
2022-05-29 15:44:35 +00:00
|
|
|
tr_training_list_timerange = []
|
|
|
|
tr_backtesting_list_timerange = []
|
2022-05-03 08:14:17 +00:00
|
|
|
first = True
|
2022-07-22 15:37:51 +00:00
|
|
|
|
2022-05-03 08:14:17 +00:00
|
|
|
while True:
|
2022-05-04 15:42:34 +00:00
|
|
|
if not first:
|
2022-08-09 13:30:25 +00:00
|
|
|
timerange_train.startts = timerange_train.startts + int(bt_period)
|
2022-07-10 10:34:09 +00:00
|
|
|
timerange_train.stopts = timerange_train.startts + train_period_days
|
2022-05-03 08:14:17 +00:00
|
|
|
|
|
|
|
first = False
|
2022-09-06 17:46:58 +00:00
|
|
|
start = datetime.fromtimestamp(timerange_train.startts, tz=timezone.utc)
|
|
|
|
stop = datetime.fromtimestamp(timerange_train.stopts, tz=timezone.utc)
|
2022-05-04 15:42:34 +00:00
|
|
|
tr_training_list.append(start.strftime("%Y%m%d") + "-" + stop.strftime("%Y%m%d"))
|
2022-05-29 15:44:35 +00:00
|
|
|
tr_training_list_timerange.append(copy.deepcopy(timerange_train))
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
# associated backtest period
|
2022-05-06 11:06:54 +00:00
|
|
|
|
|
|
|
timerange_backtest.startts = timerange_train.stopts
|
|
|
|
|
2022-08-09 13:30:25 +00:00
|
|
|
timerange_backtest.stopts = timerange_backtest.startts + int(bt_period)
|
2022-05-06 13:10:11 +00:00
|
|
|
|
2022-05-06 10:54:49 +00:00
|
|
|
if timerange_backtest.stopts > config_timerange.stopts:
|
|
|
|
timerange_backtest.stopts = config_timerange.stopts
|
|
|
|
|
2022-09-06 17:46:58 +00:00
|
|
|
start = datetime.fromtimestamp(timerange_backtest.startts, tz=timezone.utc)
|
|
|
|
stop = datetime.fromtimestamp(timerange_backtest.stopts, tz=timezone.utc)
|
2022-05-04 15:42:34 +00:00
|
|
|
tr_backtesting_list.append(start.strftime("%Y%m%d") + "-" + stop.strftime("%Y%m%d"))
|
2022-05-29 15:44:35 +00:00
|
|
|
tr_backtesting_list_timerange.append(copy.deepcopy(timerange_backtest))
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-05-06 10:54:49 +00:00
|
|
|
# ensure we are predicting on exactly same amount of data as requested by user defined
|
|
|
|
# --timerange
|
|
|
|
if timerange_backtest.stopts == config_timerange.stopts:
|
|
|
|
break
|
|
|
|
|
2022-05-29 15:44:35 +00:00
|
|
|
# print(tr_training_list, tr_backtesting_list)
|
|
|
|
return tr_training_list_timerange, tr_backtesting_list_timerange
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-09-25 13:35:55 +00:00
|
|
|
def split_timerange_live_models(
|
|
|
|
self
|
|
|
|
) -> Tuple[list, list]:
|
|
|
|
|
|
|
|
tr_backtesting_list_timerange = []
|
2022-09-29 04:48:38 +00:00
|
|
|
asset = self.pair.split("/")[0]
|
|
|
|
if asset not in self.backtest_live_models_data["assets_end_dates"]:
|
2022-09-27 03:14:12 +00:00
|
|
|
raise OperationalException(
|
|
|
|
f"Model not available for pair {self.pair}. "
|
|
|
|
"Please, try again after removing this pair from the configuration file."
|
|
|
|
)
|
2022-09-29 04:48:38 +00:00
|
|
|
asset_data = self.backtest_live_models_data["assets_end_dates"][asset]
|
2022-09-25 13:35:55 +00:00
|
|
|
backtesting_timerange = self.backtest_live_models_data["backtesting_timerange"]
|
2022-09-29 04:48:38 +00:00
|
|
|
model_end_dates = [x for x in asset_data]
|
2022-09-25 13:35:55 +00:00
|
|
|
model_end_dates.append(backtesting_timerange.stopts)
|
|
|
|
model_end_dates.sort()
|
|
|
|
for index, item in enumerate(model_end_dates):
|
|
|
|
if len(model_end_dates) > (index + 1):
|
|
|
|
tr_to_add = TimeRange("date", "date", item, model_end_dates[index + 1])
|
|
|
|
tr_backtesting_list_timerange.append(tr_to_add)
|
|
|
|
|
|
|
|
return tr_backtesting_list_timerange, tr_backtesting_list_timerange
|
|
|
|
|
2022-05-29 15:44:35 +00:00
|
|
|
def slice_dataframe(self, timerange: TimeRange, df: DataFrame) -> DataFrame:
|
2022-05-03 08:14:17 +00:00
|
|
|
"""
|
|
|
|
Given a full dataframe, extract the user desired window
|
2022-08-06 12:55:46 +00:00
|
|
|
:param tr: timerange string that we wish to extract from df
|
|
|
|
:param df: Dataframe containing all candles to run the entire backtest. Here
|
|
|
|
it is sliced down to just the present training period.
|
2022-05-03 08:14:17 +00:00
|
|
|
"""
|
2022-07-22 15:37:51 +00:00
|
|
|
|
2022-09-03 13:52:29 +00:00
|
|
|
start = datetime.fromtimestamp(timerange.startts, tz=timezone.utc)
|
|
|
|
stop = datetime.fromtimestamp(timerange.stopts, tz=timezone.utc)
|
2022-05-04 15:42:34 +00:00
|
|
|
df = df.loc[df["date"] >= start, :]
|
2022-09-04 22:07:05 +00:00
|
|
|
if not self.live:
|
|
|
|
df = df.loc[df["date"] < stop, :]
|
2022-05-03 08:14:17 +00:00
|
|
|
|
|
|
|
return df
|
|
|
|
|
|
|
|
def principal_component_analysis(self) -> None:
|
|
|
|
"""
|
|
|
|
Performs Principal Component Analysis on the data for dimensionality reduction
|
|
|
|
and outlier detection (see self.remove_outliers())
|
|
|
|
No parameters or returns, it acts on the data_dictionary held by the DataHandler.
|
|
|
|
"""
|
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
from sklearn.decomposition import PCA # avoid importing if we dont need it
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-09-01 19:51:33 +00:00
|
|
|
pca = PCA(0.999)
|
2022-05-04 15:42:34 +00:00
|
|
|
pca = pca.fit(self.data_dictionary["train_features"])
|
2022-09-01 19:51:33 +00:00
|
|
|
n_keep_components = pca.n_components_
|
2022-05-04 15:42:34 +00:00
|
|
|
self.data["n_kept_components"] = n_keep_components
|
2022-09-01 19:51:33 +00:00
|
|
|
n_components = self.data_dictionary["train_features"].shape[1]
|
2022-05-05 12:37:37 +00:00
|
|
|
logger.info("reduced feature dimension by %s", n_components - n_keep_components)
|
2022-09-01 19:51:33 +00:00
|
|
|
logger.info("explained variance %f", np.sum(pca.explained_variance_ratio_))
|
2022-05-04 15:42:34 +00:00
|
|
|
|
2022-09-01 19:51:33 +00:00
|
|
|
train_components = pca.transform(self.data_dictionary["train_features"])
|
2022-05-04 15:42:34 +00:00
|
|
|
self.data_dictionary["train_features"] = pd.DataFrame(
|
|
|
|
data=train_components,
|
|
|
|
columns=["PC" + str(i) for i in range(0, n_keep_components)],
|
|
|
|
index=self.data_dictionary["train_features"].index,
|
|
|
|
)
|
2022-09-01 19:51:33 +00:00
|
|
|
# normalsing transformed training features
|
2022-09-03 14:54:30 +00:00
|
|
|
self.data_dictionary["train_features"] = self.normalize_single_dataframe(
|
2022-09-01 19:51:33 +00:00
|
|
|
self.data_dictionary["train_features"])
|
2022-05-04 15:42:34 +00:00
|
|
|
|
2022-05-28 09:11:41 +00:00
|
|
|
# keeping a copy of the non-transformed features so we can check for errors during
|
|
|
|
# model load from disk
|
2022-07-03 08:59:38 +00:00
|
|
|
self.data["training_features_list_raw"] = copy.deepcopy(self.training_features_list)
|
2022-05-28 09:11:41 +00:00
|
|
|
self.training_features_list = self.data_dictionary["train_features"].columns
|
|
|
|
|
2022-07-25 17:40:13 +00:00
|
|
|
if self.freqai_config.get('data_split_parameters', {}).get('test_size', 0.1) != 0:
|
2022-09-01 19:51:33 +00:00
|
|
|
test_components = pca.transform(self.data_dictionary["test_features"])
|
2022-07-25 17:40:13 +00:00
|
|
|
self.data_dictionary["test_features"] = pd.DataFrame(
|
|
|
|
data=test_components,
|
|
|
|
columns=["PC" + str(i) for i in range(0, n_keep_components)],
|
|
|
|
index=self.data_dictionary["test_features"].index,
|
|
|
|
)
|
2022-09-01 19:51:33 +00:00
|
|
|
# normalise transformed test feature to transformed training features
|
|
|
|
self.data_dictionary["test_features"] = self.normalize_data_from_metadata(
|
|
|
|
self.data_dictionary["test_features"])
|
2022-05-04 15:42:34 +00:00
|
|
|
|
|
|
|
self.data["n_kept_components"] = n_keep_components
|
2022-09-01 19:51:33 +00:00
|
|
|
self.pca = pca
|
2022-05-04 15:42:34 +00:00
|
|
|
|
2022-07-03 08:59:38 +00:00
|
|
|
logger.info(f"PCA reduced total features from {n_components} to {n_keep_components}")
|
2022-05-22 15:51:49 +00:00
|
|
|
|
2022-05-23 19:05:05 +00:00
|
|
|
if not self.data_path.is_dir():
|
|
|
|
self.data_path.mkdir(parents=True, exist_ok=True)
|
2022-05-03 08:14:17 +00:00
|
|
|
|
|
|
|
return None
|
|
|
|
|
2022-05-23 10:07:09 +00:00
|
|
|
def pca_transform(self, filtered_dataframe: DataFrame) -> None:
|
2022-06-03 13:19:46 +00:00
|
|
|
"""
|
|
|
|
Use an existing pca transform to transform data into components
|
2022-10-10 12:15:30 +00:00
|
|
|
:param filtered_dataframe: DataFrame = the cleaned dataframe
|
2022-06-03 13:19:46 +00:00
|
|
|
"""
|
2022-05-23 10:07:09 +00:00
|
|
|
pca_components = self.pca.transform(filtered_dataframe)
|
|
|
|
self.data_dictionary["prediction_features"] = pd.DataFrame(
|
|
|
|
data=pca_components,
|
|
|
|
columns=["PC" + str(i) for i in range(0, self.data["n_kept_components"])],
|
|
|
|
index=filtered_dataframe.index,
|
|
|
|
)
|
2022-09-01 19:51:33 +00:00
|
|
|
# normalise transformed predictions to transformed training features
|
|
|
|
self.data_dictionary["prediction_features"] = self.normalize_data_from_metadata(
|
|
|
|
self.data_dictionary["prediction_features"])
|
2022-05-23 10:07:09 +00:00
|
|
|
|
2022-05-03 08:14:17 +00:00
|
|
|
def compute_distances(self) -> float:
|
2022-06-03 13:19:46 +00:00
|
|
|
"""
|
|
|
|
Compute distances between each training point and every other training
|
|
|
|
point. This metric defines the neighborhood of trained data and is used
|
|
|
|
for prediction confidence in the Dissimilarity Index
|
|
|
|
"""
|
2022-08-06 05:25:59 +00:00
|
|
|
# logger.info("computing average mean distance for all training points")
|
2022-08-04 15:00:59 +00:00
|
|
|
pairwise = pairwise_distances(
|
|
|
|
self.data_dictionary["train_features"], n_jobs=self.thread_count)
|
2022-08-17 13:09:57 +00:00
|
|
|
# remove the diagonal distances which are itself distances ~0
|
|
|
|
np.fill_diagonal(pairwise, np.NaN)
|
|
|
|
pairwise = pairwise.reshape(-1, 1)
|
|
|
|
avg_mean_dist = pairwise[~np.isnan(pairwise)].mean()
|
2022-08-19 10:39:08 +00:00
|
|
|
|
2022-05-03 08:14:17 +00:00
|
|
|
return avg_mean_dist
|
|
|
|
|
2022-08-27 10:44:55 +00:00
|
|
|
def get_outlier_percentage(self, dropped_pts: npt.NDArray) -> float:
|
2022-08-26 21:05:07 +00:00
|
|
|
"""
|
|
|
|
Check if more than X% of points werer dropped during outlier detection.
|
|
|
|
"""
|
|
|
|
outlier_protection_pct = self.freqai_config["feature_parameters"].get(
|
|
|
|
"outlier_protection_percentage", 30)
|
2022-08-28 10:11:29 +00:00
|
|
|
outlier_pct = (dropped_pts.sum() / len(dropped_pts)) * 100
|
2022-08-26 21:05:07 +00:00
|
|
|
if outlier_pct >= outlier_protection_pct:
|
|
|
|
return outlier_pct
|
|
|
|
else:
|
|
|
|
return 0.0
|
|
|
|
|
2022-05-22 15:51:49 +00:00
|
|
|
def use_SVM_to_remove_outliers(self, predict: bool) -> None:
|
2022-06-03 13:19:46 +00:00
|
|
|
"""
|
|
|
|
Build/inference a Support Vector Machine to detect outliers
|
|
|
|
in training data and prediction
|
2022-10-10 12:15:30 +00:00
|
|
|
:param predict: bool = If true, inference an existing SVM model, else construct one
|
2022-06-03 13:19:46 +00:00
|
|
|
"""
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-07-12 16:09:17 +00:00
|
|
|
if self.keras:
|
2022-07-21 09:25:28 +00:00
|
|
|
logger.warning(
|
|
|
|
"SVM outlier removal not currently supported for Keras based models. "
|
|
|
|
"Skipping user requested function."
|
|
|
|
)
|
2022-07-12 16:09:17 +00:00
|
|
|
if predict:
|
|
|
|
self.do_predict = np.ones(len(self.data_dictionary["prediction_features"]))
|
|
|
|
return
|
|
|
|
|
2022-05-03 08:14:17 +00:00
|
|
|
if predict:
|
2022-07-29 06:12:50 +00:00
|
|
|
if not self.svm_model:
|
|
|
|
logger.warning("No svm model available for outlier removal")
|
|
|
|
return
|
2022-05-22 15:51:49 +00:00
|
|
|
y_pred = self.svm_model.predict(self.data_dictionary["prediction_features"])
|
|
|
|
do_predict = np.where(y_pred == -1, 0, y_pred)
|
2022-05-04 15:42:34 +00:00
|
|
|
|
2022-06-26 17:02:17 +00:00
|
|
|
if (len(do_predict) - do_predict.sum()) > 0:
|
2022-08-06 12:55:46 +00:00
|
|
|
logger.info(f"SVM tossed {len(do_predict) - do_predict.sum()} predictions.")
|
2022-05-03 08:14:17 +00:00
|
|
|
self.do_predict += do_predict
|
|
|
|
self.do_predict -= 1
|
|
|
|
|
|
|
|
else:
|
2022-05-22 15:51:49 +00:00
|
|
|
# use SGDOneClassSVM to increase speed?
|
2022-07-30 11:40:05 +00:00
|
|
|
svm_params = self.freqai_config["feature_parameters"].get(
|
|
|
|
"svm_params", {"shuffle": False, "nu": 0.1})
|
|
|
|
self.svm_model = linear_model.SGDOneClassSVM(**svm_params).fit(
|
2022-07-03 08:59:38 +00:00
|
|
|
self.data_dictionary["train_features"]
|
|
|
|
)
|
2022-05-22 15:51:49 +00:00
|
|
|
y_pred = self.svm_model.predict(self.data_dictionary["train_features"])
|
2022-08-28 10:11:29 +00:00
|
|
|
kept_points = np.where(y_pred == -1, 0, y_pred)
|
2022-05-22 15:51:49 +00:00
|
|
|
# keep_index = np.where(y_pred == 1)
|
2022-08-28 10:11:29 +00:00
|
|
|
outlier_pct = self.get_outlier_percentage(1 - kept_points)
|
|
|
|
if outlier_pct:
|
2022-08-26 21:05:07 +00:00
|
|
|
logger.warning(
|
2022-08-28 10:11:29 +00:00
|
|
|
f"SVM detected {outlier_pct:.2f}% of the points as outliers. "
|
2022-08-26 21:05:07 +00:00
|
|
|
f"Keeping original dataset."
|
|
|
|
)
|
2022-08-30 10:54:39 +00:00
|
|
|
self.svm_model = None
|
2022-08-26 21:05:07 +00:00
|
|
|
return
|
|
|
|
|
2022-07-03 08:59:38 +00:00
|
|
|
self.data_dictionary["train_features"] = self.data_dictionary["train_features"][
|
|
|
|
(y_pred == 1)
|
|
|
|
]
|
|
|
|
self.data_dictionary["train_labels"] = self.data_dictionary["train_labels"][
|
|
|
|
(y_pred == 1)
|
|
|
|
]
|
|
|
|
self.data_dictionary["train_weights"] = self.data_dictionary["train_weights"][
|
|
|
|
(y_pred == 1)
|
|
|
|
]
|
2022-05-22 15:51:49 +00:00
|
|
|
|
|
|
|
logger.info(
|
2022-08-28 10:11:29 +00:00
|
|
|
f"SVM tossed {len(y_pred) - kept_points.sum()}"
|
2022-08-06 05:25:59 +00:00
|
|
|
f" train points from {len(y_pred)} total points."
|
2022-05-22 15:51:49 +00:00
|
|
|
)
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-05-22 15:51:49 +00:00
|
|
|
# same for test data
|
2022-08-06 12:55:46 +00:00
|
|
|
# TODO: This (and the part above) could be refactored into a separate function
|
|
|
|
# to reduce code duplication
|
2022-07-29 06:12:50 +00:00
|
|
|
if self.freqai_config['data_split_parameters'].get('test_size', 0.1) != 0:
|
2022-07-25 17:40:13 +00:00
|
|
|
y_pred = self.svm_model.predict(self.data_dictionary["test_features"])
|
2022-08-28 10:11:29 +00:00
|
|
|
kept_points = np.where(y_pred == -1, 0, y_pred)
|
2022-07-25 17:40:13 +00:00
|
|
|
self.data_dictionary["test_features"] = self.data_dictionary["test_features"][
|
|
|
|
(y_pred == 1)
|
|
|
|
]
|
|
|
|
self.data_dictionary["test_labels"] = self.data_dictionary["test_labels"][(
|
|
|
|
y_pred == 1)]
|
|
|
|
self.data_dictionary["test_weights"] = self.data_dictionary["test_weights"][
|
|
|
|
(y_pred == 1)
|
|
|
|
]
|
2022-05-22 15:51:49 +00:00
|
|
|
|
|
|
|
logger.info(
|
2022-08-28 10:11:29 +00:00
|
|
|
f"SVM tossed {len(y_pred) - kept_points.sum()}"
|
2022-08-06 05:25:59 +00:00
|
|
|
f" test points from {len(y_pred)} total points."
|
2022-05-22 15:51:49 +00:00
|
|
|
)
|
2022-05-03 08:14:17 +00:00
|
|
|
|
|
|
|
return
|
|
|
|
|
2022-08-04 15:41:58 +00:00
|
|
|
def use_DBSCAN_to_remove_outliers(self, predict: bool, eps=None) -> None:
|
2022-08-04 10:14:56 +00:00
|
|
|
"""
|
|
|
|
Use DBSCAN to cluster training data and remove "noisy" data (read outliers).
|
|
|
|
User controls this via the config param `DBSCAN_outlier_pct` which indicates the
|
|
|
|
pct of training data that they want to be considered outliers.
|
2022-10-10 12:15:30 +00:00
|
|
|
:param predict: bool = If False (training), iterate to find the best hyper parameters
|
|
|
|
to match user requested outlier percent target.
|
|
|
|
If True (prediction), use the parameters determined from
|
|
|
|
the previous training to estimate if the current prediction point
|
|
|
|
is an outlier.
|
2022-08-04 10:14:56 +00:00
|
|
|
"""
|
|
|
|
|
|
|
|
if predict:
|
2022-08-30 10:54:39 +00:00
|
|
|
if not self.data['DBSCAN_eps']:
|
|
|
|
return
|
2022-08-04 10:14:56 +00:00
|
|
|
train_ft_df = self.data_dictionary['train_features']
|
|
|
|
pred_ft_df = self.data_dictionary['prediction_features']
|
|
|
|
num_preds = len(pred_ft_df)
|
|
|
|
df = pd.concat([train_ft_df, pred_ft_df], axis=0, ignore_index=True)
|
2022-08-06 12:55:46 +00:00
|
|
|
clustering = DBSCAN(eps=self.data['DBSCAN_eps'],
|
|
|
|
min_samples=self.data['DBSCAN_min_samples'],
|
|
|
|
n_jobs=self.thread_count
|
2022-08-04 10:14:56 +00:00
|
|
|
).fit(df)
|
|
|
|
do_predict = np.where(clustering.labels_[-num_preds:] == -1, 0, 1)
|
|
|
|
|
|
|
|
if (len(do_predict) - do_predict.sum()) > 0:
|
2022-08-06 12:55:46 +00:00
|
|
|
logger.info(f"DBSCAN tossed {len(do_predict) - do_predict.sum()} predictions")
|
2022-08-04 10:14:56 +00:00
|
|
|
self.do_predict += do_predict
|
|
|
|
self.do_predict -= 1
|
|
|
|
|
|
|
|
else:
|
2022-08-04 15:00:59 +00:00
|
|
|
|
2022-08-22 17:57:20 +00:00
|
|
|
def normalise_distances(distances):
|
|
|
|
normalised_distances = (distances - distances.min()) / \
|
|
|
|
(distances.max() - distances.min())
|
|
|
|
return normalised_distances
|
|
|
|
|
|
|
|
def rotate_point(origin, point, angle):
|
|
|
|
# rotate a point counterclockwise by a given angle (in radians)
|
|
|
|
# around a given origin
|
|
|
|
x = origin[0] + cos(angle) * (point[0] - origin[0]) - \
|
|
|
|
sin(angle) * (point[1] - origin[1])
|
|
|
|
y = origin[1] + sin(angle) * (point[0] - origin[0]) + \
|
|
|
|
cos(angle) * (point[1] - origin[1])
|
|
|
|
return (x, y)
|
|
|
|
|
2022-08-26 16:57:27 +00:00
|
|
|
MinPts = int(len(self.data_dictionary['train_features'].index) * 0.25)
|
2022-08-28 10:56:39 +00:00
|
|
|
# measure pairwise distances to nearest neighbours
|
2022-08-05 19:26:39 +00:00
|
|
|
neighbors = NearestNeighbors(
|
|
|
|
n_neighbors=MinPts, n_jobs=self.thread_count)
|
|
|
|
neighbors_fit = neighbors.fit(self.data_dictionary['train_features'])
|
|
|
|
distances, _ = neighbors_fit.kneighbors(self.data_dictionary['train_features'])
|
2022-08-22 17:57:20 +00:00
|
|
|
distances = np.sort(distances, axis=0).mean(axis=1)
|
|
|
|
|
|
|
|
normalised_distances = normalise_distances(distances)
|
|
|
|
x_range = np.linspace(0, 1, len(distances))
|
|
|
|
line = np.linspace(normalised_distances[0],
|
|
|
|
normalised_distances[-1], len(normalised_distances))
|
|
|
|
deflection = np.abs(normalised_distances - line)
|
|
|
|
max_deflection_loc = np.where(deflection == deflection.max())[0][0]
|
|
|
|
origin = x_range[max_deflection_loc], line[max_deflection_loc]
|
|
|
|
point = x_range[max_deflection_loc], normalised_distances[max_deflection_loc]
|
|
|
|
rot_angle = np.pi / 4
|
|
|
|
elbow_loc = rotate_point(origin, point, rot_angle)
|
|
|
|
|
|
|
|
epsilon = elbow_loc[1] * (distances[-1] - distances[0]) + distances[0]
|
2022-08-05 19:26:39 +00:00
|
|
|
|
|
|
|
clustering = DBSCAN(eps=epsilon, min_samples=MinPts,
|
|
|
|
n_jobs=int(self.thread_count)).fit(
|
|
|
|
self.data_dictionary['train_features']
|
|
|
|
)
|
2022-08-04 15:00:59 +00:00
|
|
|
|
2022-08-22 17:57:20 +00:00
|
|
|
logger.info(f'DBSCAN found eps of {epsilon:.2f}.')
|
2022-08-04 10:14:56 +00:00
|
|
|
|
2022-08-04 15:41:58 +00:00
|
|
|
self.data['DBSCAN_eps'] = epsilon
|
2022-08-04 10:14:56 +00:00
|
|
|
self.data['DBSCAN_min_samples'] = MinPts
|
|
|
|
dropped_points = np.where(clustering.labels_ == -1, 1, 0)
|
|
|
|
|
2022-08-28 10:11:29 +00:00
|
|
|
outlier_pct = self.get_outlier_percentage(dropped_points)
|
|
|
|
if outlier_pct:
|
2022-08-26 21:05:07 +00:00
|
|
|
logger.warning(
|
2022-08-28 10:11:29 +00:00
|
|
|
f"DBSCAN detected {outlier_pct:.2f}% of the points as outliers. "
|
2022-08-26 21:05:07 +00:00
|
|
|
f"Keeping original dataset."
|
|
|
|
)
|
2022-08-30 10:54:39 +00:00
|
|
|
self.data['DBSCAN_eps'] = 0
|
2022-08-26 21:05:07 +00:00
|
|
|
return
|
|
|
|
|
2022-08-04 10:14:56 +00:00
|
|
|
self.data_dictionary['train_features'] = self.data_dictionary['train_features'][
|
|
|
|
(clustering.labels_ != -1)
|
|
|
|
]
|
|
|
|
self.data_dictionary["train_labels"] = self.data_dictionary["train_labels"][
|
|
|
|
(clustering.labels_ != -1)
|
|
|
|
]
|
|
|
|
self.data_dictionary["train_weights"] = self.data_dictionary["train_weights"][
|
|
|
|
(clustering.labels_ != -1)
|
|
|
|
]
|
|
|
|
|
|
|
|
logger.info(
|
|
|
|
f"DBSCAN tossed {dropped_points.sum()}"
|
|
|
|
f" train points from {len(clustering.labels_)}"
|
|
|
|
)
|
|
|
|
|
|
|
|
return
|
2022-08-18 17:15:29 +00:00
|
|
|
|
|
|
|
def compute_inlier_metric(self, set_='train') -> None:
|
2022-08-18 12:44:49 +00:00
|
|
|
"""
|
2022-08-18 17:15:29 +00:00
|
|
|
Compute inlier metric from backwards distance distributions.
|
|
|
|
This metric defines how well features from a timepoint fit
|
2022-08-18 12:44:49 +00:00
|
|
|
into previous timepoints.
|
|
|
|
"""
|
|
|
|
|
2022-09-22 16:11:50 +00:00
|
|
|
def normalise(dataframe: DataFrame, key: str) -> DataFrame:
|
|
|
|
if set_ == 'train':
|
|
|
|
min_value = dataframe.min()
|
|
|
|
max_value = dataframe.max()
|
|
|
|
self.data[f'{key}_min'] = min_value
|
|
|
|
self.data[f'{key}_max'] = max_value
|
|
|
|
else:
|
|
|
|
min_value = self.data[f'{key}_min']
|
|
|
|
max_value = self.data[f'{key}_max']
|
|
|
|
return (dataframe - min_value) / (max_value - min_value)
|
|
|
|
|
2022-08-18 17:15:29 +00:00
|
|
|
no_prev_pts = self.freqai_config["feature_parameters"]["inlier_metric_window"]
|
|
|
|
|
|
|
|
if set_ == 'train':
|
|
|
|
compute_df = copy.deepcopy(self.data_dictionary['train_features'])
|
|
|
|
elif set_ == 'test':
|
|
|
|
compute_df = copy.deepcopy(self.data_dictionary['test_features'])
|
|
|
|
else:
|
|
|
|
compute_df = copy.deepcopy(self.data_dictionary['prediction_features'])
|
|
|
|
|
|
|
|
compute_df_reindexed = compute_df.reindex(
|
|
|
|
index=np.flip(compute_df.index)
|
2022-08-18 12:44:49 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
pairwise = pd.DataFrame(
|
|
|
|
np.triu(
|
2022-08-18 17:15:29 +00:00
|
|
|
pairwise_distances(compute_df_reindexed, n_jobs=self.thread_count)
|
2022-08-18 12:44:49 +00:00
|
|
|
),
|
2022-08-18 17:15:29 +00:00
|
|
|
columns=compute_df_reindexed.index,
|
|
|
|
index=compute_df_reindexed.index
|
2022-08-18 12:44:49 +00:00
|
|
|
)
|
|
|
|
pairwise = pairwise.round(5)
|
|
|
|
|
|
|
|
column_labels = [
|
2022-08-18 17:15:29 +00:00
|
|
|
'{}{}'.format('d', i) for i in range(1, no_prev_pts + 1)
|
2022-08-18 12:44:49 +00:00
|
|
|
]
|
|
|
|
distances = pd.DataFrame(
|
2022-08-18 17:15:29 +00:00
|
|
|
columns=column_labels, index=compute_df.index
|
2022-08-18 12:44:49 +00:00
|
|
|
)
|
2022-08-18 17:15:29 +00:00
|
|
|
|
|
|
|
for index in compute_df.index[no_prev_pts:]:
|
2022-08-18 12:44:49 +00:00
|
|
|
current_row = pairwise.loc[[index]]
|
|
|
|
current_row_no_zeros = current_row.loc[
|
2022-08-18 17:15:29 +00:00
|
|
|
:, (current_row != 0).any(axis=0)
|
2022-08-18 12:44:49 +00:00
|
|
|
]
|
|
|
|
distances.loc[[index]] = current_row_no_zeros.iloc[
|
2022-08-18 17:15:29 +00:00
|
|
|
:, :no_prev_pts
|
2022-08-18 12:44:49 +00:00
|
|
|
]
|
|
|
|
distances = distances.replace([np.inf, -np.inf], np.nan)
|
2022-09-30 13:43:05 +00:00
|
|
|
drop_index = pd.isnull(distances).any(axis=1)
|
2022-08-18 17:15:29 +00:00
|
|
|
distances = distances[drop_index == 0]
|
2022-08-18 12:44:49 +00:00
|
|
|
|
|
|
|
inliers = pd.DataFrame(index=distances.index)
|
|
|
|
for key in distances.keys():
|
|
|
|
current_distances = distances[key].dropna()
|
2022-09-22 16:11:50 +00:00
|
|
|
current_distances = normalise(current_distances, key)
|
|
|
|
if set_ == 'train':
|
|
|
|
fit_params = stats.weibull_min.fit(current_distances)
|
|
|
|
self.data[f'{key}_fit_params'] = fit_params
|
|
|
|
else:
|
|
|
|
fit_params = self.data[f'{key}_fit_params']
|
2022-09-06 17:46:58 +00:00
|
|
|
quantiles = stats.weibull_min.cdf(current_distances, *fit_params)
|
2022-08-30 18:41:37 +00:00
|
|
|
|
2022-08-18 12:44:49 +00:00
|
|
|
df_inlier = pd.DataFrame(
|
2022-08-30 18:41:37 +00:00
|
|
|
{key: quantiles}, index=distances.index
|
2022-08-18 12:44:49 +00:00
|
|
|
)
|
|
|
|
inliers = pd.concat(
|
|
|
|
[inliers, df_inlier], axis=1
|
|
|
|
)
|
|
|
|
|
2022-08-18 17:15:29 +00:00
|
|
|
inlier_metric = pd.DataFrame(
|
|
|
|
data=inliers.sum(axis=1) / no_prev_pts,
|
2022-09-25 09:18:10 +00:00
|
|
|
columns=['%-inlier_metric'],
|
2022-08-18 17:15:29 +00:00
|
|
|
index=compute_df.index
|
2022-08-18 12:44:49 +00:00
|
|
|
)
|
|
|
|
|
2022-09-06 17:46:58 +00:00
|
|
|
inlier_metric = (2 * (inlier_metric - inlier_metric.min()) /
|
|
|
|
(inlier_metric.max() - inlier_metric.min()) - 1)
|
2022-08-18 17:15:29 +00:00
|
|
|
|
|
|
|
if set_ in ('train', 'test'):
|
|
|
|
inlier_metric = inlier_metric.iloc[no_prev_pts:]
|
|
|
|
compute_df = compute_df.iloc[no_prev_pts:]
|
|
|
|
self.remove_beginning_points_from_data_dict(set_, no_prev_pts)
|
|
|
|
self.data_dictionary[f'{set_}_features'] = pd.concat(
|
|
|
|
[compute_df, inlier_metric], axis=1)
|
|
|
|
else:
|
|
|
|
self.data_dictionary['prediction_features'] = pd.concat(
|
|
|
|
[compute_df, inlier_metric], axis=1)
|
|
|
|
self.data_dictionary['prediction_features'].fillna(0, inplace=True)
|
2022-08-18 12:44:49 +00:00
|
|
|
|
2022-09-03 13:52:29 +00:00
|
|
|
logger.info('Inlier metric computed and added to features.')
|
|
|
|
|
2022-08-18 12:44:49 +00:00
|
|
|
return None
|
2022-08-04 10:14:56 +00:00
|
|
|
|
2022-08-18 17:15:29 +00:00
|
|
|
def remove_beginning_points_from_data_dict(self, set_='train', no_prev_pts: int = 10):
|
|
|
|
features = self.data_dictionary[f'{set_}_features']
|
|
|
|
weights = self.data_dictionary[f'{set_}_weights']
|
|
|
|
labels = self.data_dictionary[f'{set_}_labels']
|
|
|
|
self.data_dictionary[f'{set_}_weights'] = weights[no_prev_pts:]
|
|
|
|
self.data_dictionary[f'{set_}_features'] = features.iloc[no_prev_pts:]
|
|
|
|
self.data_dictionary[f'{set_}_labels'] = labels.iloc[no_prev_pts:]
|
|
|
|
|
2022-08-19 16:35:24 +00:00
|
|
|
def add_noise_to_training_features(self) -> None:
|
|
|
|
"""
|
|
|
|
Add noise to train features to reduce the risk of overfitting.
|
|
|
|
"""
|
|
|
|
mu = 0 # no shift
|
|
|
|
sigma = self.freqai_config["feature_parameters"]["noise_standard_deviation"]
|
|
|
|
compute_df = self.data_dictionary['train_features']
|
|
|
|
noise = np.random.normal(mu, sigma, [compute_df.shape[0], compute_df.shape[1]])
|
|
|
|
self.data_dictionary['train_features'] += noise
|
|
|
|
return
|
|
|
|
|
2022-07-02 16:09:38 +00:00
|
|
|
def find_features(self, dataframe: DataFrame) -> None:
|
2022-06-03 13:19:46 +00:00
|
|
|
"""
|
|
|
|
Find features in the strategy provided dataframe
|
2022-08-06 12:55:46 +00:00
|
|
|
:param dataframe: DataFrame = strategy provided dataframe
|
|
|
|
:return:
|
2022-06-03 13:19:46 +00:00
|
|
|
features: list = the features to be used for training/prediction
|
|
|
|
"""
|
2022-05-17 16:15:03 +00:00
|
|
|
column_names = dataframe.columns
|
2022-07-03 08:59:38 +00:00
|
|
|
features = [c for c in column_names if "%" in c]
|
2022-09-30 22:22:05 +00:00
|
|
|
|
2022-05-28 09:11:41 +00:00
|
|
|
if not features:
|
|
|
|
raise OperationalException("Could not find any features!")
|
2022-07-02 16:09:38 +00:00
|
|
|
|
|
|
|
self.training_features_list = features
|
2022-09-25 09:18:10 +00:00
|
|
|
|
|
|
|
def find_labels(self, dataframe: DataFrame) -> None:
|
|
|
|
column_names = dataframe.columns
|
|
|
|
labels = [c for c in column_names if "&" in c]
|
2022-07-02 16:09:38 +00:00
|
|
|
self.label_list = labels
|
2022-05-03 08:14:17 +00:00
|
|
|
|
|
|
|
def check_if_pred_in_training_spaces(self) -> None:
|
|
|
|
"""
|
2022-05-04 15:42:34 +00:00
|
|
|
Compares the distance from each prediction point to each training data
|
2022-05-03 08:14:17 +00:00
|
|
|
point. It uses this information to estimate a Dissimilarity Index (DI)
|
2022-05-04 15:42:34 +00:00
|
|
|
and avoid making predictions on any points that are too far away
|
|
|
|
from the training data set.
|
2022-05-03 08:14:17 +00:00
|
|
|
"""
|
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
distance = pairwise_distances(
|
|
|
|
self.data_dictionary["train_features"],
|
|
|
|
self.data_dictionary["prediction_features"],
|
2022-08-04 10:14:56 +00:00
|
|
|
n_jobs=self.thread_count,
|
2022-05-04 15:42:34 +00:00
|
|
|
)
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-06-03 13:19:46 +00:00
|
|
|
self.DI_values = distance.min(axis=0) / self.data["avg_mean_dist"]
|
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
do_predict = np.where(
|
2022-07-29 06:12:50 +00:00
|
|
|
self.DI_values < self.freqai_config["feature_parameters"]["DI_threshold"],
|
2022-05-04 15:42:34 +00:00
|
|
|
1,
|
|
|
|
0,
|
|
|
|
)
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-06-26 17:02:17 +00:00
|
|
|
if (len(do_predict) - do_predict.sum()) > 0:
|
|
|
|
logger.info(
|
2022-08-22 17:57:20 +00:00
|
|
|
f"DI tossed {len(do_predict) - do_predict.sum()} predictions for "
|
2022-08-30 10:54:39 +00:00
|
|
|
"being too far from training data."
|
2022-06-26 17:02:17 +00:00
|
|
|
)
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-05-04 15:42:34 +00:00
|
|
|
self.do_predict += do_predict
|
2022-05-03 08:14:17 +00:00
|
|
|
self.do_predict -= 1
|
2022-05-04 15:42:34 +00:00
|
|
|
|
2022-05-06 14:20:52 +00:00
|
|
|
def set_weights_higher_recent(self, num_weights: int) -> npt.ArrayLike:
|
2022-05-03 08:14:17 +00:00
|
|
|
"""
|
|
|
|
Set weights so that recent data is more heavily weighted during
|
|
|
|
training than older data.
|
|
|
|
"""
|
2022-07-19 15:49:18 +00:00
|
|
|
wfactor = self.config["freqai"]["feature_parameters"]["weight_factor"]
|
2022-07-21 09:25:28 +00:00
|
|
|
weights = np.exp(-np.arange(num_weights) / (wfactor * num_weights))[::-1]
|
2022-05-03 08:14:17 +00:00
|
|
|
return weights
|
|
|
|
|
2022-08-31 14:23:48 +00:00
|
|
|
def get_predictions_to_append(self, predictions: DataFrame,
|
|
|
|
do_predict: npt.ArrayLike) -> DataFrame:
|
2022-05-03 08:14:17 +00:00
|
|
|
"""
|
2022-08-31 14:23:48 +00:00
|
|
|
Get backtest prediction from current backtest period
|
2022-05-03 08:14:17 +00:00
|
|
|
"""
|
|
|
|
|
2022-07-29 15:27:35 +00:00
|
|
|
append_df = DataFrame()
|
2022-08-06 11:51:19 +00:00
|
|
|
for label in predictions.columns:
|
2022-07-29 15:27:35 +00:00
|
|
|
append_df[label] = predictions[label]
|
2022-08-06 11:51:19 +00:00
|
|
|
if append_df[label].dtype == object:
|
|
|
|
continue
|
2022-07-29 15:27:35 +00:00
|
|
|
append_df[f"{label}_mean"] = self.data["labels_mean"][label]
|
|
|
|
append_df[f"{label}_std"] = self.data["labels_std"][label]
|
2022-05-03 08:14:17 +00:00
|
|
|
|
2022-07-29 15:27:35 +00:00
|
|
|
append_df["do_predict"] = do_predict
|
2022-07-29 06:12:50 +00:00
|
|
|
if self.freqai_config["feature_parameters"].get("DI_threshold", 0) > 0:
|
2022-07-29 15:27:35 +00:00
|
|
|
append_df["DI_values"] = self.DI_values
|
2022-07-03 15:34:44 +00:00
|
|
|
|
2022-08-31 14:23:48 +00:00
|
|
|
return append_df
|
|
|
|
|
|
|
|
def append_predictions(self, append_df: DataFrame) -> None:
|
|
|
|
"""
|
|
|
|
Append backtest prediction from current backtest period to all previous periods
|
|
|
|
"""
|
|
|
|
|
2022-07-03 15:34:44 +00:00
|
|
|
if self.full_df.empty:
|
2022-07-29 15:27:35 +00:00
|
|
|
self.full_df = append_df
|
2022-07-03 15:34:44 +00:00
|
|
|
else:
|
2022-07-29 15:27:35 +00:00
|
|
|
self.full_df = pd.concat([self.full_df, append_df], axis=0)
|
2022-07-03 15:34:44 +00:00
|
|
|
|
|
|
|
def fill_predictions(self, dataframe):
|
2022-05-03 08:14:17 +00:00
|
|
|
"""
|
|
|
|
Back fill values to before the backtesting range so that the dataframe matches size
|
|
|
|
when it goes back to the strategy. These rows are not included in the backtest.
|
|
|
|
"""
|
|
|
|
|
2022-07-03 15:34:44 +00:00
|
|
|
len_filler = len(dataframe) - len(self.full_df.index) # startup_candle_count
|
|
|
|
filler_df = pd.DataFrame(
|
|
|
|
np.zeros((len_filler, len(self.full_df.columns))), columns=self.full_df.columns
|
|
|
|
)
|
|
|
|
|
|
|
|
self.full_df = pd.concat([filler_df, self.full_df], axis=0, ignore_index=True)
|
|
|
|
|
|
|
|
to_keep = [col for col in dataframe.columns if not col.startswith("&")]
|
|
|
|
self.return_dataframe = pd.concat([dataframe[to_keep], self.full_df], axis=1)
|
|
|
|
self.full_df = DataFrame()
|
2022-05-03 08:14:17 +00:00
|
|
|
|
|
|
|
return
|
2022-05-04 15:42:34 +00:00
|
|
|
|
2022-07-10 10:34:09 +00:00
|
|
|
def create_fulltimerange(self, backtest_tr: str, backtest_period_days: int) -> str:
|
2022-07-19 14:16:44 +00:00
|
|
|
|
|
|
|
if not isinstance(backtest_period_days, int):
|
2022-07-21 09:25:28 +00:00
|
|
|
raise OperationalException("backtest_period_days must be an integer")
|
2022-07-19 14:16:44 +00:00
|
|
|
|
|
|
|
if backtest_period_days < 0:
|
2022-07-21 09:25:28 +00:00
|
|
|
raise OperationalException("backtest_period_days must be positive")
|
2022-07-19 14:16:44 +00:00
|
|
|
|
2022-05-05 13:35:51 +00:00
|
|
|
backtest_timerange = TimeRange.parse_timerange(backtest_tr)
|
|
|
|
|
2022-05-17 17:50:06 +00:00
|
|
|
if backtest_timerange.stopts == 0:
|
2022-07-21 11:02:52 +00:00
|
|
|
# typically open ended time ranges do work, however, there are some edge cases where
|
2022-08-06 12:55:46 +00:00
|
|
|
# it does not. accommodating these kinds of edge cases just to allow open-ended
|
2022-07-21 11:02:52 +00:00
|
|
|
# timerange is not high enough priority to warrant the effort. It is safer for now
|
|
|
|
# to simply ask user to add their end date
|
|
|
|
raise OperationalException("FreqAI backtesting does not allow open ended timeranges. "
|
|
|
|
"Please indicate the end date of your desired backtesting. "
|
|
|
|
"timerange.")
|
|
|
|
# backtest_timerange.stopts = int(
|
2022-09-03 13:52:29 +00:00
|
|
|
# datetime.now(tz=timezone.utc).timestamp()
|
2022-07-21 11:02:52 +00:00
|
|
|
# )
|
2022-05-17 17:50:06 +00:00
|
|
|
|
2022-07-21 09:25:28 +00:00
|
|
|
backtest_timerange.startts = (
|
|
|
|
backtest_timerange.startts - backtest_period_days * SECONDS_IN_DAY
|
|
|
|
)
|
2022-09-06 17:46:58 +00:00
|
|
|
start = datetime.fromtimestamp(backtest_timerange.startts, tz=timezone.utc)
|
|
|
|
stop = datetime.fromtimestamp(backtest_timerange.stopts, tz=timezone.utc)
|
2022-05-05 13:35:51 +00:00
|
|
|
full_timerange = start.strftime("%Y%m%d") + "-" + stop.strftime("%Y%m%d")
|
2022-05-09 13:25:00 +00:00
|
|
|
config_path = Path(self.config["config_files"][0])
|
|
|
|
|
2022-05-05 13:35:51 +00:00
|
|
|
if not self.full_path.is_dir():
|
|
|
|
self.full_path.mkdir(parents=True, exist_ok=True)
|
|
|
|
shutil.copy(
|
2022-05-15 13:26:09 +00:00
|
|
|
config_path.resolve(),
|
2022-05-09 13:25:00 +00:00
|
|
|
Path(self.full_path / config_path.parts[-1]),
|
2022-05-05 13:35:51 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
return full_timerange
|
|
|
|
|
2022-06-17 12:55:40 +00:00
|
|
|
def check_if_model_expired(self, trained_timestamp: int) -> bool:
|
2022-06-17 14:16:23 +00:00
|
|
|
"""
|
|
|
|
A model age checker to determine if the model is trustworthy based on user defined
|
|
|
|
`expiration_hours` in the configuration file.
|
2022-08-06 12:55:46 +00:00
|
|
|
:param trained_timestamp: int = The time of training for the most recent model.
|
|
|
|
:return:
|
|
|
|
bool = If the model is expired or not.
|
2022-06-17 14:16:23 +00:00
|
|
|
"""
|
2022-09-03 13:52:29 +00:00
|
|
|
time = datetime.now(tz=timezone.utc).timestamp()
|
2022-06-17 12:55:40 +00:00
|
|
|
elapsed_time = (time - trained_timestamp) / 3600 # hours
|
2022-07-03 08:59:38 +00:00
|
|
|
max_time = self.freqai_config.get("expiration_hours", 0)
|
2022-06-21 06:12:51 +00:00
|
|
|
if max_time > 0:
|
|
|
|
return elapsed_time > max_time
|
2022-06-28 13:12:25 +00:00
|
|
|
else:
|
2022-06-21 06:12:51 +00:00
|
|
|
return False
|
2022-06-17 12:55:40 +00:00
|
|
|
|
2022-07-03 08:59:38 +00:00
|
|
|
def check_if_new_training_required(
|
|
|
|
self, trained_timestamp: int
|
|
|
|
) -> Tuple[bool, TimeRange, TimeRange]:
|
2022-05-09 13:25:00 +00:00
|
|
|
|
2022-09-03 13:52:29 +00:00
|
|
|
time = datetime.now(tz=timezone.utc).timestamp()
|
2022-05-28 09:38:57 +00:00
|
|
|
trained_timerange = TimeRange()
|
2022-05-31 16:42:27 +00:00
|
|
|
data_load_timerange = TimeRange()
|
|
|
|
|
2022-08-09 13:30:25 +00:00
|
|
|
timeframes = self.freqai_config["feature_parameters"].get("include_timeframes")
|
2022-05-31 16:42:27 +00:00
|
|
|
|
2022-08-09 13:30:25 +00:00
|
|
|
max_tf_seconds = 0
|
|
|
|
for tf in timeframes:
|
|
|
|
secs = timeframe_to_seconds(tf)
|
|
|
|
if secs > max_tf_seconds:
|
|
|
|
max_tf_seconds = secs
|
|
|
|
|
|
|
|
# We notice that users like to use exotic indicators where
|
|
|
|
# they do not know the required timeperiod. Here we include a factor
|
|
|
|
# of safety by multiplying the user considered "max" by 2.
|
2022-08-22 16:19:07 +00:00
|
|
|
max_period = self.config.get('startup_candle_count', 20) * 2
|
2022-08-09 13:30:25 +00:00
|
|
|
additional_seconds = max_period * max_tf_seconds
|
2022-06-03 13:19:46 +00:00
|
|
|
|
2022-05-23 19:05:05 +00:00
|
|
|
if trained_timestamp != 0:
|
2022-07-10 10:34:09 +00:00
|
|
|
elapsed_time = (time - trained_timestamp) / SECONDS_IN_HOUR
|
|
|
|
retrain = elapsed_time > self.freqai_config.get("live_retrain_hours", 0)
|
2022-05-22 22:06:26 +00:00
|
|
|
if retrain:
|
2022-07-03 08:59:38 +00:00
|
|
|
trained_timerange.startts = int(
|
2022-07-10 10:34:09 +00:00
|
|
|
time - self.freqai_config.get("train_period_days", 0) * SECONDS_IN_DAY
|
2022-07-03 08:59:38 +00:00
|
|
|
)
|
2022-05-23 19:05:05 +00:00
|
|
|
trained_timerange.stopts = int(time)
|
2022-05-31 16:42:27 +00:00
|
|
|
# we want to load/populate indicators on more data than we plan to train on so
|
|
|
|
# because most of the indicators have a rolling timeperiod, and are thus NaNs
|
|
|
|
# unless they have data further back in time before the start of the train period
|
2022-07-03 08:59:38 +00:00
|
|
|
data_load_timerange.startts = int(
|
|
|
|
time
|
2022-07-10 10:34:09 +00:00
|
|
|
- self.freqai_config.get("train_period_days", 0) * SECONDS_IN_DAY
|
2022-07-03 08:59:38 +00:00
|
|
|
- additional_seconds
|
|
|
|
)
|
2022-05-31 16:42:27 +00:00
|
|
|
data_load_timerange.stopts = int(time)
|
2022-05-22 22:06:26 +00:00
|
|
|
else: # user passed no live_trained_timerange in config
|
2022-07-03 08:59:38 +00:00
|
|
|
trained_timerange.startts = int(
|
2022-08-06 12:55:46 +00:00
|
|
|
time - self.freqai_config.get("train_period_days", 0) * SECONDS_IN_DAY
|
2022-07-03 08:59:38 +00:00
|
|
|
)
|
2022-05-22 15:51:49 +00:00
|
|
|
trained_timerange.stopts = int(time)
|
2022-05-31 16:42:27 +00:00
|
|
|
|
2022-07-03 08:59:38 +00:00
|
|
|
data_load_timerange.startts = int(
|
|
|
|
time
|
2022-07-10 10:34:09 +00:00
|
|
|
- self.freqai_config.get("train_period_days", 0) * SECONDS_IN_DAY
|
2022-07-03 08:59:38 +00:00
|
|
|
- additional_seconds
|
|
|
|
)
|
2022-05-31 16:42:27 +00:00
|
|
|
data_load_timerange.stopts = int(time)
|
2022-05-22 15:51:49 +00:00
|
|
|
retrain = True
|
2022-05-09 13:25:00 +00:00
|
|
|
|
2022-05-31 16:42:27 +00:00
|
|
|
return retrain, trained_timerange, data_load_timerange
|
2022-05-09 13:25:00 +00:00
|
|
|
|
2022-09-25 13:35:55 +00:00
|
|
|
def set_new_model_names(self, pair: str, timestamp_id: int):
|
2022-05-23 19:05:05 +00:00
|
|
|
|
2022-06-08 04:14:01 +00:00
|
|
|
coin, _ = pair.split("/")
|
2022-07-03 08:59:38 +00:00
|
|
|
self.data_path = Path(
|
|
|
|
self.full_path
|
2022-09-25 13:35:55 +00:00
|
|
|
/ f"sub-train-{pair.split('/')[0]}_{timestamp_id}"
|
2022-07-03 08:59:38 +00:00
|
|
|
)
|
2022-05-23 19:05:05 +00:00
|
|
|
|
2022-09-25 13:35:55 +00:00
|
|
|
self.model_filename = f"cb_{coin.lower()}_{timestamp_id}"
|
2022-05-25 12:40:32 +00:00
|
|
|
|
2022-06-03 13:19:46 +00:00
|
|
|
def set_all_pairs(self) -> None:
|
|
|
|
|
2022-07-21 09:25:28 +00:00
|
|
|
self.all_pairs = copy.deepcopy(
|
2022-07-29 06:12:50 +00:00
|
|
|
self.freqai_config["feature_parameters"].get("include_corr_pairlist", [])
|
2022-07-21 09:25:28 +00:00
|
|
|
)
|
2022-07-03 08:59:38 +00:00
|
|
|
for pair in self.config.get("exchange", "").get("pair_whitelist"):
|
2022-06-03 13:19:46 +00:00
|
|
|
if pair not in self.all_pairs:
|
|
|
|
self.all_pairs.append(pair)
|
|
|
|
|
2022-07-03 08:59:38 +00:00
|
|
|
def use_strategy_to_populate_indicators(
|
2022-07-21 10:24:22 +00:00
|
|
|
self,
|
|
|
|
strategy: IStrategy,
|
|
|
|
corr_dataframes: dict = {},
|
|
|
|
base_dataframes: dict = {},
|
|
|
|
pair: str = "",
|
|
|
|
prediction_dataframe: DataFrame = pd.DataFrame(),
|
2022-07-03 08:59:38 +00:00
|
|
|
) -> DataFrame:
|
2022-06-03 13:19:46 +00:00
|
|
|
"""
|
2022-10-10 12:15:30 +00:00
|
|
|
Use the user defined strategy for populating indicators during retrain
|
|
|
|
:param strategy: IStrategy = user defined strategy object
|
|
|
|
:param corr_dataframes: dict = dict containing the informative pair dataframes
|
|
|
|
(for user defined timeframes)
|
|
|
|
:param base_dataframes: dict = dict containing the current pair dataframes
|
|
|
|
(for user defined timeframes)
|
|
|
|
:param metadata: dict = strategy furnished pair metadata
|
2022-06-03 13:19:46 +00:00
|
|
|
:returns:
|
|
|
|
dataframe: DataFrame = dataframe containing populated indicators
|
|
|
|
"""
|
2022-07-21 10:24:22 +00:00
|
|
|
|
|
|
|
# for prediction dataframe creation, we let dataprovider handle everything in the strategy
|
|
|
|
# so we create empty dictionaries, which allows us to pass None to
|
|
|
|
# `populate_any_indicators()`. Signaling we want the dp to give us the live dataframe.
|
2022-07-29 06:12:50 +00:00
|
|
|
tfs = self.freqai_config["feature_parameters"].get("include_timeframes")
|
|
|
|
pairs = self.freqai_config["feature_parameters"].get("include_corr_pairlist", [])
|
2022-07-21 10:24:22 +00:00
|
|
|
if not prediction_dataframe.empty:
|
|
|
|
dataframe = prediction_dataframe.copy()
|
|
|
|
for tf in tfs:
|
|
|
|
base_dataframes[tf] = None
|
|
|
|
for p in pairs:
|
|
|
|
if p not in corr_dataframes:
|
|
|
|
corr_dataframes[p] = {}
|
|
|
|
corr_dataframes[p][tf] = None
|
|
|
|
else:
|
|
|
|
dataframe = base_dataframes[self.config["timeframe"]].copy()
|
|
|
|
|
2022-07-31 15:05:29 +00:00
|
|
|
sgi = False
|
2022-07-21 10:24:22 +00:00
|
|
|
for tf in tfs:
|
2022-07-31 15:05:29 +00:00
|
|
|
if tf == tfs[-1]:
|
|
|
|
sgi = True # doing this last allows user to use all tf raw prices in labels
|
2022-08-14 19:42:55 +00:00
|
|
|
dataframe = strategy.populate_any_indicators(
|
|
|
|
pair,
|
|
|
|
dataframe.copy(),
|
|
|
|
tf,
|
|
|
|
informative=base_dataframes[tf],
|
|
|
|
set_generalized_indicators=sgi
|
|
|
|
)
|
|
|
|
if pairs:
|
|
|
|
for i in pairs:
|
|
|
|
if pair in i:
|
|
|
|
continue # dont repeat anything from whitelist
|
|
|
|
dataframe = strategy.populate_any_indicators(
|
|
|
|
i,
|
|
|
|
dataframe.copy(),
|
|
|
|
tf,
|
|
|
|
informative=corr_dataframes[i][tf]
|
|
|
|
)
|
2022-05-09 13:25:00 +00:00
|
|
|
|
2022-08-09 15:31:38 +00:00
|
|
|
self.get_unique_classes_from_labels(dataframe)
|
|
|
|
|
2022-05-09 13:25:00 +00:00
|
|
|
return dataframe
|
|
|
|
|
2022-05-26 19:07:50 +00:00
|
|
|
def fit_labels(self) -> None:
|
2022-06-03 13:19:46 +00:00
|
|
|
"""
|
|
|
|
Fit the labels with a gaussian distribution
|
|
|
|
"""
|
2022-05-26 19:07:50 +00:00
|
|
|
import scipy as spy
|
|
|
|
|
2022-07-03 08:59:38 +00:00
|
|
|
self.data["labels_mean"], self.data["labels_std"] = {}, {}
|
2022-08-06 11:51:19 +00:00
|
|
|
for label in self.data_dictionary["train_labels"].columns:
|
2022-07-09 08:13:33 +00:00
|
|
|
if self.data_dictionary["train_labels"][label].dtype == object:
|
|
|
|
continue
|
2022-07-02 16:09:38 +00:00
|
|
|
f = spy.stats.norm.fit(self.data_dictionary["train_labels"][label])
|
|
|
|
self.data["labels_mean"][label], self.data["labels_std"][label] = f[0], f[1]
|
2022-05-26 19:07:50 +00:00
|
|
|
|
2022-08-10 13:16:50 +00:00
|
|
|
# incase targets are classifications
|
|
|
|
for label in self.unique_class_list:
|
|
|
|
self.data["labels_mean"][label], self.data["labels_std"][label] = 0, 0
|
|
|
|
|
2022-05-26 19:07:50 +00:00
|
|
|
return
|
|
|
|
|
2022-07-22 10:17:15 +00:00
|
|
|
def remove_features_from_df(self, dataframe: DataFrame) -> DataFrame:
|
|
|
|
"""
|
|
|
|
Remove the features from the dataframe before returning it to strategy. This keeps it
|
|
|
|
compact for Frequi purposes.
|
|
|
|
"""
|
|
|
|
to_keep = [
|
|
|
|
col for col in dataframe.columns if not col.startswith("%") or col.startswith("%%")
|
|
|
|
]
|
|
|
|
return dataframe[to_keep]
|
2022-08-09 15:31:38 +00:00
|
|
|
|
|
|
|
def get_unique_classes_from_labels(self, dataframe: DataFrame) -> None:
|
|
|
|
|
2022-09-25 09:18:10 +00:00
|
|
|
# self.find_features(dataframe)
|
|
|
|
self.find_labels(dataframe)
|
2022-08-09 15:31:38 +00:00
|
|
|
|
|
|
|
for key in self.label_list:
|
|
|
|
if dataframe[key].dtype == object:
|
|
|
|
self.unique_classes[key] = dataframe[key].dropna().unique()
|
2022-08-10 13:16:50 +00:00
|
|
|
|
|
|
|
if self.unique_classes:
|
|
|
|
for label in self.unique_classes:
|
|
|
|
self.unique_class_list += list(self.unique_classes[label])
|
2022-08-31 14:23:48 +00:00
|
|
|
|
|
|
|
def save_backtesting_prediction(
|
2022-09-01 10:09:23 +00:00
|
|
|
self, append_df: DataFrame
|
2022-08-31 14:23:48 +00:00
|
|
|
) -> None:
|
|
|
|
"""
|
|
|
|
Save prediction dataframe from backtesting to h5 file format
|
2022-09-01 10:09:23 +00:00
|
|
|
:param append_df: dataframe for backtesting period
|
2022-08-31 14:23:48 +00:00
|
|
|
"""
|
2022-09-03 12:00:01 +00:00
|
|
|
full_predictions_folder = Path(self.full_path / self.backtest_predictions_folder)
|
2022-09-01 10:09:23 +00:00
|
|
|
if not full_predictions_folder.is_dir():
|
|
|
|
full_predictions_folder.mkdir(parents=True, exist_ok=True)
|
2022-08-31 14:23:48 +00:00
|
|
|
|
2022-09-01 10:09:23 +00:00
|
|
|
append_df.to_hdf(self.backtesting_results_path, key='append_df', mode='w')
|
2022-08-31 18:36:29 +00:00
|
|
|
|
|
|
|
def get_backtesting_prediction(
|
2022-09-01 10:09:23 +00:00
|
|
|
self
|
2022-08-31 18:36:29 +00:00
|
|
|
) -> DataFrame:
|
2022-08-31 14:23:48 +00:00
|
|
|
"""
|
2022-09-01 10:09:23 +00:00
|
|
|
Get prediction dataframe from h5 file format
|
2022-08-31 14:23:48 +00:00
|
|
|
"""
|
2022-09-01 10:09:23 +00:00
|
|
|
append_df = pd.read_hdf(self.backtesting_results_path)
|
2022-08-31 14:23:48 +00:00
|
|
|
return append_df
|
2022-09-03 12:00:01 +00:00
|
|
|
|
|
|
|
def check_if_backtest_prediction_exists(
|
|
|
|
self
|
|
|
|
) -> bool:
|
|
|
|
"""
|
|
|
|
Check if a backtesting prediction already exists
|
|
|
|
:param dk: FreqaiDataKitchen
|
|
|
|
:return:
|
|
|
|
:boolean: whether the prediction file exists or not.
|
|
|
|
"""
|
|
|
|
path_to_predictionfile = Path(self.full_path /
|
|
|
|
self.backtest_predictions_folder /
|
|
|
|
f"{self.model_filename}_prediction.h5")
|
|
|
|
self.backtesting_results_path = path_to_predictionfile
|
|
|
|
|
|
|
|
file_exists = path_to_predictionfile.is_file()
|
|
|
|
if file_exists:
|
|
|
|
logger.info(f"Found backtesting prediction file at {path_to_predictionfile}")
|
|
|
|
else:
|
|
|
|
logger.info(
|
|
|
|
f"Could not find backtesting prediction file at {path_to_predictionfile}"
|
|
|
|
)
|
|
|
|
return file_exists
|
2022-09-24 15:28:52 +00:00
|
|
|
|
2022-09-26 02:14:00 +00:00
|
|
|
def set_timerange_from_ready_models(self):
|
2022-09-24 16:01:53 +00:00
|
|
|
backtesting_timerange, \
|
2022-09-29 04:48:38 +00:00
|
|
|
assets_end_dates = (
|
|
|
|
freqai_util.get_timerange_and_assets_end_dates_from_ready_models(self.full_path))
|
|
|
|
|
2022-09-24 16:01:53 +00:00
|
|
|
self.backtest_live_models_data = {
|
|
|
|
"backtesting_timerange": backtesting_timerange,
|
2022-09-29 04:48:38 +00:00
|
|
|
"assets_end_dates": assets_end_dates
|
2022-09-24 16:01:53 +00:00
|
|
|
}
|
|
|
|
return
|