2017-11-18 07:34:32 +00:00
|
|
|
"""
|
2018-12-12 18:57:25 +00:00
|
|
|
Functions to convert data from one format to another
|
2017-11-18 07:34:32 +00:00
|
|
|
"""
|
2020-04-01 05:23:43 +00:00
|
|
|
import itertools
|
2018-03-25 19:37:14 +00:00
|
|
|
import logging
|
2019-12-25 14:47:04 +00:00
|
|
|
from datetime import datetime, timezone
|
2020-04-01 05:23:43 +00:00
|
|
|
from operator import itemgetter
|
2020-03-31 18:12:01 +00:00
|
|
|
from typing import Any, Dict, List
|
2019-04-09 09:27:35 +00:00
|
|
|
|
2018-08-05 04:41:06 +00:00
|
|
|
import pandas as pd
|
2018-03-02 15:22:00 +00:00
|
|
|
from pandas import DataFrame, to_datetime
|
2018-03-17 21:44:47 +00:00
|
|
|
|
2020-04-01 05:23:43 +00:00
|
|
|
from freqtrade.constants import (DEFAULT_DATAFRAME_COLUMNS,
|
|
|
|
DEFAULT_TRADES_COLUMNS)
|
2018-12-30 15:07:47 +00:00
|
|
|
|
2018-03-25 19:37:14 +00:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
2020-03-08 10:35:31 +00:00
|
|
|
def ohlcv_to_dataframe(ohlcv: list, timeframe: str, pair: str, *,
|
|
|
|
fill_missing: bool = True, drop_incomplete: bool = True) -> DataFrame:
|
2018-07-10 10:04:37 +00:00
|
|
|
"""
|
2020-03-08 10:35:31 +00:00
|
|
|
Converts a list with candle (OHLCV) data (in format returned by ccxt.fetch_ohlcv)
|
|
|
|
to a Dataframe
|
|
|
|
:param ohlcv: list with candle (OHLCV) data, as returned by exchange.async_get_candle_history
|
2019-11-02 19:19:13 +00:00
|
|
|
:param timeframe: timeframe (e.g. 5m). Used to fill up eventual missing data
|
2019-06-15 11:47:20 +00:00
|
|
|
:param pair: Pair this data is for (used to warn if fillup was necessary)
|
2018-12-31 18:42:14 +00:00
|
|
|
:param fill_missing: fill up missing candles with 0 candles
|
|
|
|
(see ohlcv_fill_up_missing_data for details)
|
2019-06-09 12:35:58 +00:00
|
|
|
:param drop_incomplete: Drop the last candle of the dataframe, assuming it's incomplete
|
2018-07-10 10:04:37 +00:00
|
|
|
:return: DataFrame
|
|
|
|
"""
|
2020-03-08 10:35:31 +00:00
|
|
|
logger.debug(f"Converting candle (OHLCV) data to dataframe for pair {pair}.")
|
2019-12-26 18:52:08 +00:00
|
|
|
cols = DEFAULT_DATAFRAME_COLUMNS
|
2020-03-08 10:35:31 +00:00
|
|
|
df = DataFrame(ohlcv, columns=cols)
|
2018-07-10 10:04:37 +00:00
|
|
|
|
2020-03-08 10:35:31 +00:00
|
|
|
df['date'] = to_datetime(df['date'], unit='ms', utc=True, infer_datetime_format=True)
|
2018-07-10 10:04:37 +00:00
|
|
|
|
2020-03-08 10:35:31 +00:00
|
|
|
# Some exchanges return int values for Volume and even for OHLC.
|
2019-02-10 19:23:00 +00:00
|
|
|
# Convert them since TA-LIB indicators used in the strategy assume floats
|
2019-02-10 19:13:40 +00:00
|
|
|
# and fail with exception...
|
2020-03-08 10:35:31 +00:00
|
|
|
df = df.astype(dtype={'open': 'float', 'high': 'float', 'low': 'float', 'close': 'float',
|
|
|
|
'volume': 'float'})
|
|
|
|
return clean_ohlcv_dataframe(df, timeframe, pair,
|
2019-12-25 12:24:09 +00:00
|
|
|
fill_missing=fill_missing,
|
|
|
|
drop_incomplete=drop_incomplete)
|
2019-02-10 18:52:33 +00:00
|
|
|
|
2019-12-25 12:24:09 +00:00
|
|
|
|
|
|
|
def clean_ohlcv_dataframe(data: DataFrame, timeframe: str, pair: str, *,
|
|
|
|
fill_missing: bool = True,
|
|
|
|
drop_incomplete: bool = True) -> DataFrame:
|
|
|
|
"""
|
2020-03-08 10:35:31 +00:00
|
|
|
Clense a OHLCV dataframe by
|
2019-12-25 12:24:09 +00:00
|
|
|
* Grouping it by date (removes duplicate tics)
|
|
|
|
* dropping last candles if requested
|
|
|
|
* Filling up missing data (if requested)
|
2020-03-08 10:35:31 +00:00
|
|
|
:param data: DataFrame containing candle (OHLCV) data.
|
2019-12-25 12:24:09 +00:00
|
|
|
:param timeframe: timeframe (e.g. 5m). Used to fill up eventual missing data
|
|
|
|
:param pair: Pair this data is for (used to warn if fillup was necessary)
|
|
|
|
:param fill_missing: fill up missing candles with 0 candles
|
|
|
|
(see ohlcv_fill_up_missing_data for details)
|
|
|
|
:param drop_incomplete: Drop the last candle of the dataframe, assuming it's incomplete
|
|
|
|
:return: DataFrame
|
|
|
|
"""
|
2018-07-10 10:04:37 +00:00
|
|
|
# group by index and aggregate results to eliminate duplicate ticks
|
2019-12-25 12:24:09 +00:00
|
|
|
data = data.groupby(by='date', as_index=False, sort=True).agg({
|
2018-07-10 10:04:37 +00:00
|
|
|
'open': 'first',
|
|
|
|
'high': 'max',
|
|
|
|
'low': 'min',
|
|
|
|
'close': 'last',
|
|
|
|
'volume': 'max',
|
|
|
|
})
|
2019-06-09 12:35:58 +00:00
|
|
|
# eliminate partial candle
|
|
|
|
if drop_incomplete:
|
2019-12-25 12:24:09 +00:00
|
|
|
data.drop(data.tail(1).index, inplace=True)
|
2019-06-09 12:35:58 +00:00
|
|
|
logger.debug('Dropping last candle')
|
2018-12-31 18:13:34 +00:00
|
|
|
|
|
|
|
if fill_missing:
|
2019-12-25 12:24:09 +00:00
|
|
|
return ohlcv_fill_up_missing_data(data, timeframe, pair)
|
2018-12-31 18:13:34 +00:00
|
|
|
else:
|
2019-12-25 12:24:09 +00:00
|
|
|
return data
|
2018-08-05 04:41:06 +00:00
|
|
|
|
|
|
|
|
2019-11-02 19:19:13 +00:00
|
|
|
def ohlcv_fill_up_missing_data(dataframe: DataFrame, timeframe: str, pair: str) -> DataFrame:
|
2018-12-30 15:07:47 +00:00
|
|
|
"""
|
|
|
|
Fills up missing data with 0 volume rows,
|
|
|
|
using the previous close as price for "open", "high" "low" and "close", volume is set to 0
|
|
|
|
|
|
|
|
"""
|
2019-04-09 09:27:35 +00:00
|
|
|
from freqtrade.exchange import timeframe_to_minutes
|
|
|
|
|
2020-03-08 10:35:31 +00:00
|
|
|
ohlcv_dict = {
|
2018-12-30 15:07:47 +00:00
|
|
|
'open': 'first',
|
|
|
|
'high': 'max',
|
|
|
|
'low': 'min',
|
|
|
|
'close': 'last',
|
|
|
|
'volume': 'sum'
|
|
|
|
}
|
2020-03-08 10:35:31 +00:00
|
|
|
timeframe_minutes = timeframe_to_minutes(timeframe)
|
2018-12-30 15:07:47 +00:00
|
|
|
# Resample to create "NAN" values
|
2020-03-08 10:35:31 +00:00
|
|
|
df = dataframe.resample(f'{timeframe_minutes}min', on='date').agg(ohlcv_dict)
|
2018-12-30 15:07:47 +00:00
|
|
|
|
|
|
|
# Forwardfill close for missing columns
|
|
|
|
df['close'] = df['close'].fillna(method='ffill')
|
|
|
|
# Use close for "open, high, low"
|
|
|
|
df.loc[:, ['open', 'high', 'low']] = df[['open', 'high', 'low']].fillna(
|
|
|
|
value={'open': df['close'],
|
|
|
|
'high': df['close'],
|
|
|
|
'low': df['close'],
|
|
|
|
})
|
|
|
|
df.reset_index(inplace=True)
|
2019-06-15 11:31:14 +00:00
|
|
|
len_before = len(dataframe)
|
|
|
|
len_after = len(df)
|
|
|
|
if len_before != len_after:
|
2019-06-15 11:46:19 +00:00
|
|
|
logger.info(f"Missing data fillup for {pair}: before: {len_before} - after: {len_after}")
|
2018-12-30 15:07:47 +00:00
|
|
|
return df
|
|
|
|
|
|
|
|
|
2019-12-28 18:58:41 +00:00
|
|
|
def trim_dataframe(df: DataFrame, timerange, df_date_col: str = 'date') -> DataFrame:
|
2019-12-25 14:47:04 +00:00
|
|
|
"""
|
|
|
|
Trim dataframe based on given timerange
|
|
|
|
:param df: Dataframe to trim
|
|
|
|
:param timerange: timerange (use start and end date if available)
|
|
|
|
:param: df_date_col: Column in the dataframe to use as Date column
|
|
|
|
:return: trimmed dataframe
|
|
|
|
"""
|
|
|
|
if timerange.starttype == 'date':
|
|
|
|
start = datetime.fromtimestamp(timerange.startts, tz=timezone.utc)
|
|
|
|
df = df.loc[df[df_date_col] >= start, :]
|
|
|
|
if timerange.stoptype == 'date':
|
|
|
|
stop = datetime.fromtimestamp(timerange.stopts, tz=timezone.utc)
|
|
|
|
df = df.loc[df[df_date_col] <= stop, :]
|
|
|
|
return df
|
|
|
|
|
|
|
|
|
2018-08-05 13:08:07 +00:00
|
|
|
def order_book_to_dataframe(bids: list, asks: list) -> DataFrame:
|
2018-08-05 04:41:06 +00:00
|
|
|
"""
|
2019-12-28 09:54:10 +00:00
|
|
|
TODO: This should get a dedicated test
|
2018-08-05 04:41:06 +00:00
|
|
|
Gets order book list, returns dataframe with below format per suggested by creslin
|
|
|
|
-------------------------------------------------------------------
|
|
|
|
b_sum b_size bids asks a_size a_sum
|
|
|
|
-------------------------------------------------------------------
|
|
|
|
"""
|
|
|
|
cols = ['bids', 'b_size']
|
2018-08-05 13:08:07 +00:00
|
|
|
|
|
|
|
bids_frame = DataFrame(bids, columns=cols)
|
2018-08-05 04:41:06 +00:00
|
|
|
# add cumulative sum column
|
|
|
|
bids_frame['b_sum'] = bids_frame['b_size'].cumsum()
|
|
|
|
cols2 = ['asks', 'a_size']
|
2018-08-05 13:08:07 +00:00
|
|
|
asks_frame = DataFrame(asks, columns=cols2)
|
2018-08-05 04:41:06 +00:00
|
|
|
# add cumulative sum column
|
|
|
|
asks_frame['a_sum'] = asks_frame['a_size'].cumsum()
|
|
|
|
|
|
|
|
frame = pd.concat([bids_frame['b_sum'], bids_frame['b_size'], bids_frame['bids'],
|
|
|
|
asks_frame['asks'], asks_frame['a_size'], asks_frame['a_sum']], axis=1,
|
|
|
|
keys=['b_sum', 'b_size', 'bids', 'asks', 'a_size', 'a_sum'])
|
|
|
|
# logger.info('order book %s', frame )
|
|
|
|
return frame
|
2019-10-13 17:21:27 +00:00
|
|
|
|
|
|
|
|
2020-04-01 05:23:43 +00:00
|
|
|
def trades_remove_duplicates(trades: List[List]) -> List[List]:
|
|
|
|
"""
|
|
|
|
Removes duplicates from the trades list.
|
|
|
|
Uses itertools.groupby to avoid converting to pandas.
|
|
|
|
Tests show it as being pretty efficient on lists of 4M Lists.
|
|
|
|
:param trades: List of Lists with constants.DEFAULT_TRADES_COLUMNS as columns
|
|
|
|
:return: same format as above, but with duplicates removed
|
|
|
|
"""
|
|
|
|
return [i for i, _ in itertools.groupby(sorted(trades, key=itemgetter(0)))]
|
|
|
|
|
|
|
|
|
2020-03-31 18:12:01 +00:00
|
|
|
def trades_dict_to_list(trades: List[Dict]) -> List[List]:
|
|
|
|
"""
|
|
|
|
Convert fetch_trades result into a List (to be more memory efficient).
|
|
|
|
:param trades: List of trades, as returned by ccxt.fetch_trades.
|
|
|
|
:return: List of Lists, with constants.DEFAULT_TRADES_COLUMNS as columns
|
|
|
|
"""
|
2020-04-17 04:22:25 +00:00
|
|
|
return [[t[col] for col in DEFAULT_TRADES_COLUMNS] for t in trades]
|
2020-03-31 18:12:01 +00:00
|
|
|
|
|
|
|
|
|
|
|
def trades_to_ohlcv(trades: List, timeframe: str) -> DataFrame:
|
2019-10-13 17:21:27 +00:00
|
|
|
"""
|
2020-03-08 10:35:31 +00:00
|
|
|
Converts trades list to OHLCV list
|
2019-12-28 09:54:10 +00:00
|
|
|
TODO: This should get a dedicated test
|
2019-10-14 04:19:59 +00:00
|
|
|
:param trades: List of trades, as returned by ccxt.fetch_trades.
|
2020-03-08 10:35:31 +00:00
|
|
|
:param timeframe: Timeframe to resample data to
|
|
|
|
:return: OHLCV Dataframe.
|
2019-10-13 17:21:27 +00:00
|
|
|
"""
|
|
|
|
from freqtrade.exchange import timeframe_to_minutes
|
2020-03-08 10:35:31 +00:00
|
|
|
timeframe_minutes = timeframe_to_minutes(timeframe)
|
2020-03-31 18:46:42 +00:00
|
|
|
df = pd.DataFrame(trades, columns=DEFAULT_TRADES_COLUMNS)
|
|
|
|
df['timestamp'] = pd.to_datetime(df['timestamp'], unit='ms',
|
|
|
|
utc=True,)
|
|
|
|
df = df.set_index('timestamp')
|
2019-10-13 17:21:27 +00:00
|
|
|
|
2020-03-08 10:35:31 +00:00
|
|
|
df_new = df['price'].resample(f'{timeframe_minutes}min').ohlc()
|
|
|
|
df_new['volume'] = df['amount'].resample(f'{timeframe_minutes}min').sum()
|
2019-12-25 15:34:27 +00:00
|
|
|
df_new['date'] = df_new.index
|
2019-10-14 04:19:59 +00:00
|
|
|
# Drop 0 volume rows
|
2019-10-13 17:21:27 +00:00
|
|
|
df_new = df_new.dropna()
|
2020-06-02 16:37:08 +00:00
|
|
|
return df_new.loc[:, DEFAULT_DATAFRAME_COLUMNS]
|
2019-12-28 09:27:49 +00:00
|
|
|
|
|
|
|
|
|
|
|
def convert_trades_format(config: Dict[str, Any], convert_from: str, convert_to: str, erase: bool):
|
|
|
|
"""
|
|
|
|
Convert trades from one format to another format.
|
|
|
|
:param config: Config dictionary
|
|
|
|
:param convert_from: Source format
|
|
|
|
:param convert_to: Target format
|
|
|
|
:param erase: Erase souce data (does not apply if source and target format are identical)
|
|
|
|
"""
|
|
|
|
from freqtrade.data.history.idatahandler import get_datahandler
|
|
|
|
src = get_datahandler(config['datadir'], convert_from)
|
|
|
|
trg = get_datahandler(config['datadir'], convert_to)
|
|
|
|
|
|
|
|
if 'pairs' not in config:
|
|
|
|
config['pairs'] = src.trades_get_pairs(config['datadir'])
|
|
|
|
logger.info(f"Converting trades for {config['pairs']}")
|
|
|
|
|
|
|
|
for pair in config['pairs']:
|
|
|
|
data = src.trades_load(pair=pair)
|
|
|
|
logger.info(f"Converting {len(data)} trades for {pair}")
|
|
|
|
trg.trades_store(pair, data)
|
|
|
|
if erase and convert_from != convert_to:
|
|
|
|
logger.info(f"Deleting source Trade data for {pair}.")
|
|
|
|
src.trades_purge(pair=pair)
|
|
|
|
|
|
|
|
|
|
|
|
def convert_ohlcv_format(config: Dict[str, Any], convert_from: str, convert_to: str, erase: bool):
|
|
|
|
"""
|
2020-03-08 10:35:31 +00:00
|
|
|
Convert OHLCV from one format to another
|
2019-12-28 09:27:49 +00:00
|
|
|
:param config: Config dictionary
|
|
|
|
:param convert_from: Source format
|
|
|
|
:param convert_to: Target format
|
|
|
|
:param erase: Erase souce data (does not apply if source and target format are identical)
|
|
|
|
"""
|
|
|
|
from freqtrade.data.history.idatahandler import get_datahandler
|
|
|
|
src = get_datahandler(config['datadir'], convert_from)
|
|
|
|
trg = get_datahandler(config['datadir'], convert_to)
|
|
|
|
timeframes = config.get('timeframes', [config.get('ticker_interval')])
|
2020-03-08 10:35:31 +00:00
|
|
|
logger.info(f"Converting candle (OHLCV) for timeframe {timeframes}")
|
2019-12-28 09:27:49 +00:00
|
|
|
|
|
|
|
if 'pairs' not in config:
|
|
|
|
config['pairs'] = []
|
|
|
|
# Check timeframes or fall back to ticker_interval.
|
|
|
|
for timeframe in timeframes:
|
|
|
|
config['pairs'].extend(src.ohlcv_get_pairs(config['datadir'],
|
|
|
|
timeframe))
|
2020-03-08 10:35:31 +00:00
|
|
|
logger.info(f"Converting candle (OHLCV) data for {config['pairs']}")
|
2019-12-28 09:27:49 +00:00
|
|
|
|
|
|
|
for timeframe in timeframes:
|
|
|
|
for pair in config['pairs']:
|
|
|
|
data = src.ohlcv_load(pair=pair, timeframe=timeframe,
|
|
|
|
timerange=None,
|
|
|
|
fill_missing=False,
|
|
|
|
drop_incomplete=False,
|
|
|
|
startup_candles=0)
|
|
|
|
logger.info(f"Converting {len(data)} candles for {pair}")
|
|
|
|
trg.ohlcv_store(pair=pair, timeframe=timeframe, data=data)
|
|
|
|
if erase and convert_from != convert_to:
|
|
|
|
logger.info(f"Deleting source data for {pair} / {timeframe}")
|
|
|
|
src.ohlcv_purge(pair=pair, timeframe=timeframe)
|