use data loader, add evaluation on epoch
This commit is contained in:
@@ -1,6 +1,6 @@
|
||||
import logging
|
||||
from time import time
|
||||
from typing import Any, Dict
|
||||
from typing import Any
|
||||
|
||||
import torch
|
||||
from pandas import DataFrame
|
||||
@@ -11,7 +11,7 @@ from freqtrade.freqai.freqai_interface import IFreqaiModel
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class BasePytorchModel(IFreqaiModel):
|
||||
class BasePyTorchModel(IFreqaiModel):
|
||||
"""
|
||||
Base class for TensorFlow type models.
|
||||
User *must* inherit from this class and set fit() and predict().
|
||||
@@ -29,7 +29,6 @@ class BasePytorchModel(IFreqaiModel):
|
||||
Filter the training data and train a model to it. Train makes heavy use of the datakitchen
|
||||
for storing, saving, loading, and analyzing the data.
|
||||
:param unfiltered_df: Full dataframe for the current training period
|
||||
:param metadata: pair metadata from strategy.
|
||||
:return:
|
||||
:model: Trained model which can be used to inference (self.predict)
|
||||
"""
|
136
freqtrade/freqai/base_models/PyTorchModelTrainer.py
Normal file
136
freqtrade/freqai/base_models/PyTorchModelTrainer.py
Normal file
@@ -0,0 +1,136 @@
|
||||
import logging
|
||||
from pathlib import Path
|
||||
from typing import Dict
|
||||
|
||||
import torch
|
||||
import torch.nn as nn
|
||||
from torch.utils.data import DataLoader
|
||||
from torch.utils.data import TensorDataset
|
||||
import pandas as pd
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class PyTorchModelTrainer:
|
||||
def __init__(
|
||||
self,
|
||||
model: nn.Module,
|
||||
optimizer: nn.Module,
|
||||
criterion: nn.Module,
|
||||
device: str,
|
||||
batch_size: int,
|
||||
max_iters: int,
|
||||
eval_iters: int,
|
||||
init_model: Dict
|
||||
):
|
||||
self.model = model
|
||||
self.optimizer = optimizer
|
||||
self.criterion = criterion
|
||||
self.device = device
|
||||
self.max_iters = max_iters
|
||||
self.batch_size = batch_size
|
||||
self.eval_iters = eval_iters
|
||||
|
||||
if init_model:
|
||||
self.load_from_checkpoint(init_model)
|
||||
|
||||
def fit(self, data_dictionary: Dict[str, pd.DataFrame]):
|
||||
data_loaders_dictionary = self.create_data_loaders_dictionary(data_dictionary)
|
||||
epochs = self.calc_n_epochs(
|
||||
n_obs=len(data_dictionary['train_features']),
|
||||
batch_size=self.batch_size,
|
||||
n_iters=self.max_iters
|
||||
)
|
||||
for epoch in range(epochs):
|
||||
# evaluation
|
||||
losses = self.estimate_loss(data_loaders_dictionary, data_dictionary)
|
||||
logger.info(
|
||||
f"epoch ({epoch}/{epochs}):"
|
||||
f" train loss {losses['train']:.4f} ; test loss {losses['test']:.4f}"
|
||||
)
|
||||
# training
|
||||
for batch_data in data_loaders_dictionary['train']:
|
||||
xb, yb = batch_data
|
||||
xb = xb.to(self.device) # type: ignore
|
||||
yb = yb.to(self.device)
|
||||
yb_pred = self.model(xb)
|
||||
loss = self.criterion(yb_pred, yb)
|
||||
|
||||
self.optimizer.zero_grad(set_to_none=True)
|
||||
loss.backward()
|
||||
self.optimizer.step()
|
||||
|
||||
@torch.no_grad()
|
||||
def estimate_loss(
|
||||
self,
|
||||
data_loader_dictionary: Dict[str, DataLoader],
|
||||
data_dictionary: Dict[str, pd.DataFrame]
|
||||
) -> Dict[str, float]:
|
||||
|
||||
self.model.eval()
|
||||
epochs = self.calc_n_epochs(
|
||||
n_obs=len(data_dictionary[f'test_features']),
|
||||
batch_size=self.batch_size,
|
||||
n_iters=self.eval_iters
|
||||
)
|
||||
loss_dictionary = {}
|
||||
for split in ['train', 'test']:
|
||||
losses = torch.zeros(epochs)
|
||||
for i, batch in enumerate(data_loader_dictionary[split]):
|
||||
xb, yb = batch
|
||||
xb = xb.to(self.device)
|
||||
yb = yb.to(self.device)
|
||||
yb_pred = self.model(xb)
|
||||
loss = self.criterion(yb_pred, yb)
|
||||
losses[i] = loss.item()
|
||||
|
||||
loss_dictionary[split] = losses.mean()
|
||||
|
||||
self.model.train()
|
||||
return loss_dictionary
|
||||
|
||||
def create_data_loaders_dictionary(
|
||||
self,
|
||||
data_dictionary: Dict[str, pd.DataFrame]
|
||||
) -> Dict[str, DataLoader]:
|
||||
data_loader_dictionary = {}
|
||||
for split in ['train', 'test']:
|
||||
labels_shape = data_dictionary[f'{split}_labels'].shape
|
||||
labels_view = labels_shape[0] if labels_shape[1] == 1 else labels_shape
|
||||
dataset = TensorDataset(
|
||||
torch.from_numpy(data_dictionary[f'{split}_features'].values).float(),
|
||||
torch.from_numpy(data_dictionary[f'{split}_labels'].astype(float).values)
|
||||
.long()
|
||||
.view(labels_view)
|
||||
)
|
||||
data_loader = DataLoader(
|
||||
dataset,
|
||||
batch_size=self.batch_size,
|
||||
shuffle=True,
|
||||
drop_last=True,
|
||||
num_workers=0,
|
||||
)
|
||||
data_loader_dictionary[split] = data_loader
|
||||
|
||||
return data_loader_dictionary
|
||||
|
||||
@staticmethod
|
||||
def calc_n_epochs(n_obs: int, batch_size: int, n_iters: int) -> int:
|
||||
n_batches = n_obs // batch_size
|
||||
epochs = n_iters // n_batches
|
||||
return epochs
|
||||
|
||||
def save(self, path: Path):
|
||||
torch.save({
|
||||
'model_state_dict': self.model.state_dict(),
|
||||
'optimizer_state_dict': self.optimizer.state_dict(),
|
||||
}, path)
|
||||
|
||||
def load_from_file(self, path: Path):
|
||||
checkpoint = torch.load(path)
|
||||
return self.load_from_checkpoint(checkpoint)
|
||||
|
||||
def load_from_checkpoint(self, checkpoint: Dict):
|
||||
self.model.load_state_dict(checkpoint['model_state_dict'])
|
||||
self.optimizer.load_state_dict(checkpoint['optimizer_state_dict'])
|
||||
return self
|
@@ -1,51 +0,0 @@
|
||||
import logging
|
||||
from pathlib import Path
|
||||
from typing import Dict
|
||||
|
||||
import torch
|
||||
import torch.nn as nn
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class PytorchModelTrainer:
|
||||
def __init__(self, model: nn.Module, optimizer, init_model: Dict):
|
||||
self.model = model
|
||||
self.optimizer = optimizer
|
||||
if init_model:
|
||||
self.load_from_checkpoint(init_model)
|
||||
|
||||
def fit(self, tensor_dictionary, max_iters, batch_size):
|
||||
for iter in range(max_iters):
|
||||
|
||||
# todo add validation evaluation here
|
||||
|
||||
xb, yb = self.get_batch(tensor_dictionary, 'train', batch_size)
|
||||
logits, loss = self.model(xb, yb)
|
||||
|
||||
self.optimizer.zero_grad(set_to_none=True)
|
||||
loss.backward()
|
||||
self.optimizer.step()
|
||||
|
||||
def save(self, path):
|
||||
torch.save({
|
||||
'model_state_dict': self.model.state_dict(),
|
||||
'optimizer_state_dict': self.optimizer.state_dict(),
|
||||
}, path)
|
||||
|
||||
def load_from_file(self, path: Path):
|
||||
checkpoint = torch.load(path)
|
||||
return self.load_from_checkpoint(checkpoint)
|
||||
|
||||
def load_from_checkpoint(self, checkpoint: Dict):
|
||||
self.model.load_state_dict(checkpoint['model_state_dict'])
|
||||
self.optimizer.load_state_dict(checkpoint['optimizer_state_dict'])
|
||||
return self
|
||||
|
||||
@staticmethod
|
||||
def get_batch(tensor_dictionary: Dict, split: str, batch_size: int):
|
||||
ix = torch.randint(len(tensor_dictionary[f'{split}_labels']), (batch_size,))
|
||||
x = tensor_dictionary[f'{split}_features'][ix]
|
||||
y = tensor_dictionary[f'{split}_labels'][ix]
|
||||
return x, y
|
||||
|
Reference in New Issue
Block a user