2020-09-04 17:44:35 +00:00
|
|
|
import numpy as np
|
2020-09-28 17:43:15 +00:00
|
|
|
import pandas as pd
|
2021-01-04 12:47:16 +00:00
|
|
|
import pytest
|
2020-09-04 17:44:35 +00:00
|
|
|
|
2021-07-17 16:19:49 +00:00
|
|
|
from freqtrade.data.dataprovider import DataProvider
|
2021-12-08 14:59:20 +00:00
|
|
|
from freqtrade.enums import CandleType
|
2022-01-29 18:59:54 +00:00
|
|
|
from freqtrade.resolvers.strategy_resolver import StrategyResolver
|
2022-10-05 16:09:26 +00:00
|
|
|
from freqtrade.strategy import merge_informative_pair, stoploss_from_absolute, stoploss_from_open
|
|
|
|
from tests.conftest import generate_test_data, get_patched_exchange
|
2020-09-04 17:44:35 +00:00
|
|
|
|
|
|
|
|
2020-09-04 18:02:31 +00:00
|
|
|
def test_merge_informative_pair():
|
2020-09-04 17:44:35 +00:00
|
|
|
data = generate_test_data('15m', 40)
|
|
|
|
informative = generate_test_data('1h', 40)
|
|
|
|
|
2020-09-04 18:09:02 +00:00
|
|
|
result = merge_informative_pair(data, informative, '15m', '1h', ffill=True)
|
2020-09-04 17:44:35 +00:00
|
|
|
assert isinstance(result, pd.DataFrame)
|
|
|
|
assert len(result) == len(data)
|
|
|
|
assert 'date' in result.columns
|
|
|
|
assert result['date'].equals(data['date'])
|
|
|
|
assert 'date_1h' in result.columns
|
|
|
|
|
|
|
|
assert 'open' in result.columns
|
|
|
|
assert 'open_1h' in result.columns
|
|
|
|
assert result['open'].equals(data['open'])
|
|
|
|
|
|
|
|
assert 'close' in result.columns
|
|
|
|
assert 'close_1h' in result.columns
|
|
|
|
assert result['close'].equals(data['close'])
|
|
|
|
|
|
|
|
assert 'volume' in result.columns
|
|
|
|
assert 'volume_1h' in result.columns
|
|
|
|
assert result['volume'].equals(data['volume'])
|
|
|
|
|
2021-01-04 12:47:16 +00:00
|
|
|
# First 3 rows are empty
|
2020-09-04 17:44:35 +00:00
|
|
|
assert result.iloc[0]['date_1h'] is pd.NaT
|
|
|
|
assert result.iloc[1]['date_1h'] is pd.NaT
|
|
|
|
assert result.iloc[2]['date_1h'] is pd.NaT
|
|
|
|
# Next 4 rows contain the starting date (0:00)
|
2021-01-04 12:47:16 +00:00
|
|
|
assert result.iloc[3]['date_1h'] == result.iloc[0]['date']
|
2020-09-04 17:44:35 +00:00
|
|
|
assert result.iloc[4]['date_1h'] == result.iloc[0]['date']
|
|
|
|
assert result.iloc[5]['date_1h'] == result.iloc[0]['date']
|
|
|
|
assert result.iloc[6]['date_1h'] == result.iloc[0]['date']
|
|
|
|
# Next 4 rows contain the next Hourly date original date row 4
|
2021-01-04 12:47:16 +00:00
|
|
|
assert result.iloc[7]['date_1h'] == result.iloc[4]['date']
|
2020-09-04 17:44:35 +00:00
|
|
|
assert result.iloc[8]['date_1h'] == result.iloc[4]['date']
|
2020-09-04 18:09:02 +00:00
|
|
|
|
2022-04-23 09:25:20 +00:00
|
|
|
informative = generate_test_data('1h', 40)
|
|
|
|
result = merge_informative_pair(data, informative, '15m', '1h', ffill=False)
|
|
|
|
# First 3 rows are empty
|
|
|
|
assert result.iloc[0]['date_1h'] is pd.NaT
|
|
|
|
assert result.iloc[1]['date_1h'] is pd.NaT
|
|
|
|
assert result.iloc[2]['date_1h'] is pd.NaT
|
|
|
|
# Next 4 rows contain the starting date (0:00)
|
|
|
|
assert result.iloc[3]['date_1h'] == result.iloc[0]['date']
|
|
|
|
assert result.iloc[4]['date_1h'] is pd.NaT
|
|
|
|
assert result.iloc[5]['date_1h'] is pd.NaT
|
|
|
|
assert result.iloc[6]['date_1h'] is pd.NaT
|
|
|
|
# Next 4 rows contain the next Hourly date original date row 4
|
|
|
|
assert result.iloc[7]['date_1h'] == result.iloc[4]['date']
|
|
|
|
assert result.iloc[8]['date_1h'] is pd.NaT
|
|
|
|
|
2020-09-04 18:09:02 +00:00
|
|
|
|
|
|
|
def test_merge_informative_pair_same():
|
|
|
|
data = generate_test_data('15m', 40)
|
|
|
|
informative = generate_test_data('15m', 40)
|
|
|
|
|
|
|
|
result = merge_informative_pair(data, informative, '15m', '15m', ffill=True)
|
|
|
|
assert isinstance(result, pd.DataFrame)
|
|
|
|
assert len(result) == len(data)
|
|
|
|
assert 'date' in result.columns
|
|
|
|
assert result['date'].equals(data['date'])
|
|
|
|
assert 'date_15m' in result.columns
|
|
|
|
|
|
|
|
assert 'open' in result.columns
|
|
|
|
assert 'open_15m' in result.columns
|
|
|
|
assert result['open'].equals(data['open'])
|
|
|
|
|
|
|
|
assert 'close' in result.columns
|
|
|
|
assert 'close_15m' in result.columns
|
|
|
|
assert result['close'].equals(data['close'])
|
|
|
|
|
|
|
|
assert 'volume' in result.columns
|
|
|
|
assert 'volume_15m' in result.columns
|
|
|
|
assert result['volume'].equals(data['volume'])
|
|
|
|
|
|
|
|
# Dates match 1:1
|
|
|
|
assert result['date_15m'].equals(result['date'])
|
2021-01-04 12:47:16 +00:00
|
|
|
|
|
|
|
|
|
|
|
def test_merge_informative_pair_lower():
|
|
|
|
data = generate_test_data('1h', 40)
|
|
|
|
informative = generate_test_data('15m', 40)
|
|
|
|
|
|
|
|
with pytest.raises(ValueError, match=r"Tried to merge a faster timeframe .*"):
|
|
|
|
merge_informative_pair(data, informative, '1h', '15m', ffill=True)
|
2021-03-17 13:44:10 +00:00
|
|
|
|
|
|
|
|
2022-09-07 21:26:38 +00:00
|
|
|
def test_merge_informative_pair_suffix():
|
|
|
|
data = generate_test_data('15m', 20)
|
|
|
|
informative = generate_test_data('1h', 20)
|
|
|
|
|
|
|
|
result = merge_informative_pair(data, informative, '15m', '1h',
|
|
|
|
append_timeframe=False, suffix="suf")
|
|
|
|
|
|
|
|
assert 'date' in result.columns
|
|
|
|
assert result['date'].equals(data['date'])
|
|
|
|
assert 'date_suf' in result.columns
|
|
|
|
|
|
|
|
assert 'open_suf' in result.columns
|
|
|
|
assert 'open_1h' not in result.columns
|
|
|
|
|
|
|
|
|
|
|
|
def test_merge_informative_pair_suffix_append_timeframe():
|
|
|
|
data = generate_test_data('15m', 20)
|
|
|
|
informative = generate_test_data('1h', 20)
|
|
|
|
|
|
|
|
with pytest.raises(ValueError, match=r"You can not specify `append_timeframe` .*"):
|
|
|
|
merge_informative_pair(data, informative, '15m', '1h', suffix="suf")
|
|
|
|
|
|
|
|
|
2021-03-17 13:44:10 +00:00
|
|
|
def test_stoploss_from_open():
|
|
|
|
open_price_ranges = [
|
|
|
|
[0.01, 1.00, 30],
|
|
|
|
[1, 100, 30],
|
|
|
|
[100, 10000, 30],
|
|
|
|
]
|
2022-01-15 03:30:30 +00:00
|
|
|
# profit range for long is [-1, inf] while for shorts is [-inf, 1]
|
|
|
|
current_profit_range_dict = {'long': [-0.99, 2, 30], 'short': [-2.0, 0.99, 30]}
|
2021-03-17 13:44:10 +00:00
|
|
|
desired_stop_range = [-0.50, 0.50, 30]
|
|
|
|
|
2022-01-15 03:30:30 +00:00
|
|
|
for side, current_profit_range in current_profit_range_dict.items():
|
|
|
|
for open_range in open_price_ranges:
|
|
|
|
for open_price in np.linspace(*open_range):
|
|
|
|
for desired_stop in np.linspace(*desired_stop_range):
|
2021-03-17 13:44:10 +00:00
|
|
|
|
2022-01-15 03:30:30 +00:00
|
|
|
if side == 'long':
|
|
|
|
# -1 is not a valid current_profit, should return 1
|
|
|
|
assert stoploss_from_open(desired_stop, -1) == 1
|
2021-03-17 13:44:10 +00:00
|
|
|
else:
|
2022-01-15 03:30:30 +00:00
|
|
|
# 1 is not a valid current_profit for shorts, should return 1
|
|
|
|
assert stoploss_from_open(desired_stop, 1, True) == 1
|
|
|
|
|
|
|
|
for current_profit in np.linspace(*current_profit_range):
|
|
|
|
if side == 'long':
|
|
|
|
current_price = open_price * (1 + current_profit)
|
|
|
|
expected_stop_price = open_price * (1 + desired_stop)
|
|
|
|
stoploss = stoploss_from_open(desired_stop, current_profit)
|
|
|
|
stop_price = current_price * (1 - stoploss)
|
|
|
|
else:
|
|
|
|
current_price = open_price * (1 - current_profit)
|
|
|
|
expected_stop_price = open_price * (1 - desired_stop)
|
|
|
|
stoploss = stoploss_from_open(desired_stop, current_profit, True)
|
|
|
|
stop_price = current_price * (1 + stoploss)
|
|
|
|
|
|
|
|
assert stoploss >= 0
|
|
|
|
# Technically the formula can yield values greater than 1 for shorts
|
|
|
|
# eventhough it doesn't make sense because the position would be liquidated
|
|
|
|
if side == 'long':
|
|
|
|
assert stoploss <= 1
|
|
|
|
|
|
|
|
# there is no correct answer if the expected stop price is above
|
|
|
|
# the current price
|
2022-01-18 15:52:34 +00:00
|
|
|
if ((side == 'long' and expected_stop_price > current_price)
|
|
|
|
or (side == 'short' and expected_stop_price < current_price)):
|
2022-01-15 03:30:30 +00:00
|
|
|
assert stoploss == 0
|
|
|
|
else:
|
2022-08-31 08:18:28 +00:00
|
|
|
assert pytest.approx(stop_price) == expected_stop_price
|
2021-07-17 16:19:49 +00:00
|
|
|
|
|
|
|
|
2021-09-18 07:18:33 +00:00
|
|
|
def test_stoploss_from_absolute():
|
2022-01-24 18:15:42 +00:00
|
|
|
assert pytest.approx(stoploss_from_absolute(90, 100)) == 1 - (90 / 100)
|
|
|
|
assert pytest.approx(stoploss_from_absolute(90, 100)) == 0.1
|
|
|
|
assert pytest.approx(stoploss_from_absolute(95, 100)) == 0.05
|
|
|
|
assert pytest.approx(stoploss_from_absolute(100, 100)) == 0
|
|
|
|
assert pytest.approx(stoploss_from_absolute(110, 100)) == 0
|
|
|
|
assert pytest.approx(stoploss_from_absolute(100, 0)) == 1
|
|
|
|
assert pytest.approx(stoploss_from_absolute(0, 100)) == 1
|
|
|
|
|
|
|
|
assert pytest.approx(stoploss_from_absolute(90, 100, True)) == 0
|
|
|
|
assert pytest.approx(stoploss_from_absolute(100, 100, True)) == 0
|
2022-04-11 16:02:02 +00:00
|
|
|
assert pytest.approx(stoploss_from_absolute(110, 100, True)) == -(1 - (110 / 100))
|
2022-01-24 18:15:42 +00:00
|
|
|
assert pytest.approx(stoploss_from_absolute(110, 100, True)) == 0.1
|
|
|
|
assert pytest.approx(stoploss_from_absolute(105, 100, True)) == 0.05
|
|
|
|
assert pytest.approx(stoploss_from_absolute(100, 0, True)) == 1
|
|
|
|
assert pytest.approx(stoploss_from_absolute(0, 100, True)) == 0
|
|
|
|
assert pytest.approx(stoploss_from_absolute(100, 1, True)) == 1
|
2022-01-22 17:01:02 +00:00
|
|
|
|
2021-09-18 07:18:33 +00:00
|
|
|
|
2022-01-28 15:58:07 +00:00
|
|
|
@pytest.mark.parametrize('trading_mode', ['futures', 'spot'])
|
2022-01-29 18:59:54 +00:00
|
|
|
def test_informative_decorator(mocker, default_conf_usdt, trading_mode):
|
2022-01-28 15:58:07 +00:00
|
|
|
candle_def = CandleType.get_default(trading_mode)
|
2022-01-29 18:59:54 +00:00
|
|
|
default_conf_usdt['candle_type_def'] = candle_def
|
2021-07-17 16:19:49 +00:00
|
|
|
test_data_5m = generate_test_data('5m', 40)
|
|
|
|
test_data_30m = generate_test_data('30m', 40)
|
|
|
|
test_data_1h = generate_test_data('1h', 40)
|
|
|
|
data = {
|
2022-01-28 15:58:07 +00:00
|
|
|
('XRP/USDT', '5m', candle_def): test_data_5m,
|
|
|
|
('XRP/USDT', '30m', candle_def): test_data_30m,
|
|
|
|
('XRP/USDT', '1h', candle_def): test_data_1h,
|
|
|
|
('LTC/USDT', '5m', candle_def): test_data_5m,
|
|
|
|
('LTC/USDT', '30m', candle_def): test_data_30m,
|
|
|
|
('LTC/USDT', '1h', candle_def): test_data_1h,
|
|
|
|
('NEO/USDT', '30m', candle_def): test_data_30m,
|
|
|
|
('NEO/USDT', '5m', CandleType.SPOT): test_data_5m, # Explicit request with '' as candletype
|
|
|
|
('NEO/USDT', '15m', candle_def): test_data_5m, # Explicit request with '' as candletype
|
|
|
|
('NEO/USDT', '1h', candle_def): test_data_1h,
|
|
|
|
('ETH/USDT', '1h', candle_def): test_data_1h,
|
|
|
|
('ETH/USDT', '30m', candle_def): test_data_30m,
|
|
|
|
('ETH/BTC', '1h', CandleType.SPOT): test_data_1h, # Explicitly selected as spot
|
2021-07-17 16:19:49 +00:00
|
|
|
}
|
2022-01-29 18:59:54 +00:00
|
|
|
default_conf_usdt['strategy'] = 'InformativeDecoratorTest'
|
|
|
|
strategy = StrategyResolver.load_strategy(default_conf_usdt)
|
|
|
|
exchange = get_patched_exchange(mocker, default_conf_usdt)
|
2021-11-22 07:27:33 +00:00
|
|
|
strategy.dp = DataProvider({}, exchange, None)
|
2021-09-07 12:40:53 +00:00
|
|
|
mocker.patch.object(strategy.dp, 'current_whitelist', return_value=[
|
2021-11-22 07:27:33 +00:00
|
|
|
'XRP/USDT', 'LTC/USDT', 'NEO/USDT'
|
2021-09-07 12:40:53 +00:00
|
|
|
])
|
2021-07-17 16:19:49 +00:00
|
|
|
|
2021-09-10 06:36:52 +00:00
|
|
|
assert len(strategy._ft_informative) == 6 # Equal to number of decorators used
|
2021-12-07 06:25:00 +00:00
|
|
|
informative_pairs = [
|
2022-01-28 15:58:07 +00:00
|
|
|
('XRP/USDT', '1h', candle_def),
|
|
|
|
('LTC/USDT', '1h', candle_def),
|
|
|
|
('XRP/USDT', '30m', candle_def),
|
|
|
|
('LTC/USDT', '30m', candle_def),
|
|
|
|
('NEO/USDT', '1h', candle_def),
|
|
|
|
('NEO/USDT', '30m', candle_def),
|
|
|
|
('NEO/USDT', '5m', candle_def),
|
|
|
|
('NEO/USDT', '15m', candle_def),
|
|
|
|
('NEO/USDT', '2h', CandleType.FUTURES),
|
|
|
|
('ETH/BTC', '1h', CandleType.SPOT), # One candle remains as spot
|
|
|
|
('ETH/USDT', '30m', candle_def)]
|
2021-07-17 16:19:49 +00:00
|
|
|
for inf_pair in informative_pairs:
|
|
|
|
assert inf_pair in strategy.gather_informative_pairs()
|
|
|
|
|
2021-11-21 07:43:05 +00:00
|
|
|
def test_historic_ohlcv(pair, timeframe, candle_type):
|
2021-12-08 13:35:15 +00:00
|
|
|
return data[
|
|
|
|
(pair, timeframe or strategy.timeframe, CandleType.from_string(candle_type))].copy()
|
2022-01-28 15:58:07 +00:00
|
|
|
|
2021-07-17 16:19:49 +00:00
|
|
|
mocker.patch('freqtrade.data.dataprovider.DataProvider.historic_ohlcv',
|
|
|
|
side_effect=test_historic_ohlcv)
|
|
|
|
|
|
|
|
analyzed = strategy.advise_all_indicators(
|
2022-01-28 15:58:07 +00:00
|
|
|
{p: data[(p, strategy.timeframe, candle_def)] for p in ('XRP/USDT', 'LTC/USDT')})
|
2021-07-17 16:19:49 +00:00
|
|
|
expected_columns = [
|
|
|
|
'rsi_1h', 'rsi_30m', # Stacked informative decorators
|
2021-11-22 07:27:33 +00:00
|
|
|
'neo_usdt_rsi_1h', # NEO 1h informative
|
|
|
|
'rsi_NEO_USDT_neo_usdt_NEO/USDT_30m', # Column formatting
|
2021-07-17 16:19:49 +00:00
|
|
|
'rsi_from_callable', # Custom column formatter
|
|
|
|
'eth_btc_rsi_1h', # Quote currency not matching stake currency
|
|
|
|
'rsi', 'rsi_less', # Non-informative columns
|
|
|
|
'rsi_5m', # Manual informative dataframe
|
|
|
|
]
|
|
|
|
for _, dataframe in analyzed.items():
|
|
|
|
for col in expected_columns:
|
|
|
|
assert col in dataframe.columns
|