24x5 Stock Trading Agent to predict stock prices with Deep Learning with deployment

栏目: IT技术 · 发布时间: 4年前

内容简介:If you have followed the stock market recently, you would have noticed the wild swings due to COVID-19. It goes up a day and goes down another day, which might be easily predicted by AI. Won’t it be wonderful to have a stock trading agent with AI powers to

24x5 AI Stock Trading Agent to predict stock prices with Deep Learning with deployment

24x5 Stock Trading Agent to predict stock prices with Deep Learning with deployment

Apr 18 ·8min read

If you have followed the stock market recently, you would have noticed the wild swings due to COVID-19. It goes up a day and goes down another day, which might be easily predicted by AI. Won’t it be wonderful to have a stock trading agent with AI powers to buy and sell stocks without the need to monitor hour by hour?

So I decided to create a bot to trade. You would have seen a lot of models that read from CSV’s and create a neural network, LSTM, or Deep Reinforcement Models(DRL). However, those models end up in a sandbox environment and often tricky to use in a live environment. So I created the AI pipeline, which trades in real-time. After all, who does not want to make money in the stock market? So let’s get started. Below is the process we are going to follow to implement it.

  1. Alpaca Broker account
  2. Alpaca python package for API trading
  3. Collect data, EDA, feature engineering
  4. AI model and Training
  5. AWS Cloud to host code and get predictions
  6. Create a lambda function and API
  7. Trade stocks automatically

Alpaca Broker account:

Currently, most of the brokerage firms offer zero trading fees. However, not all brokerage firms have an API option to trade. Alpaca provides free trading with python API to trade. Once you create an account, you will have paper trading and live trading options. We can test the strategies in paper trading and implement it in live trading. It is just a key change for live trading.

Alpaca python package for API trading:

If you can have a local environment, you can install the pip package . Once installed, you can select Paper trading or Live trading.

24x5 Stock Trading Agent to predict stock prices with Deep Learning with deployment

Based on your selection, you can get the API and secret key.

24x5 Stock Trading Agent to predict stock prices with Deep Learning with deployment

Now, these keys will be used in our code.

import alpaca_trade_api as tradeapiapi = tradeapi.REST('xxxxxxxx', 'xxxxxxxxxx',base_url='https://paper-api.alpaca.markets', api_version='v2',)

Collect data and transform

Getting Data

One advantage of using Alpaca is you can get historical data from polygon API. The timeframe can be from minute, hour, day, etc. Once you create the dataframe, the chart should be something like this.

24x5 Stock Trading Agent to predict stock prices with Deep Learning with deployment

Feature Engineering

Like any data science project, we need to create features related to the dataset. Some part of the implementation was referred fromthis article. I have built around 430+ technical indicators from the above dataset. Features include momentum, trends, volatility, RSI, etc.

24x5 Stock Trading Agent to predict stock prices with Deep Learning with deployment

Features have been created for each day. It can be easily created for hourly or any other timeframe. For some models which we are going to create like LSTM, DRL we might need to use the original dataset.

Creating labels and features is where we have to create a logic to train our model. For now, I have used the logic from this paper .

However, creating logic can be altered according to your needs. When performing unsupervised learning, you don’t require to create labels.

Finally, the data needs to be scaled. Neural networks work better in scaled data. The first function will fit the scaler object using the train data and the next function is used to scale any dataset.

# scale train and test data to [-1, 1]
def transform_scale(train):
    # fit scaler
    
    print(len(train.columns))
    scaler = MinMaxScaler(feature_range=(-1, 1))
    scaler = scaler.fit(train)
    # transform train
    return scaler# scale train and test data to [-1, 1]
def scale(dataset, scaler):
    # transform train
    dataset = scaler.transform(dataset)
    print(dataset.shape)
    return dataset

Once we create the model, we have to prepare our data as a data loader. Below function will perform it.

def _get_train_data_loader(batch_size, train_data):
    print("Get train data loader.")
    
    train_X =    torch.from_numpy(train_data.drop(['labels'],axis=1).values).float()
    
    train_Y = torch.from_numpy(train_data['labels'].values).float()
    
    train_ds = torch.utils.data.TensorDataset(train_X,train_Y)return torch.utils.data.DataLoader(train_ds,shuffle=False, batch_size=batch_size)

AI model

In this section, we are going to create different types of models. However, these models might not be perfect for a time series dataset. I wanted to show how to use a deep learning model with a complete pipeline.

Fully connected Deep NN

Here we will create a fully connected deep neural network. The model itself is not fancy and I am not expecting to perform better. Also, it is not an appropriate model for time series data. I am using this model just to use all our features and for the sake of simplicity.

However, we are starting with a basic model to complete our pipeline. In the next section, I will show how to create other types of models. Our model.py looks like below one.

import torch.nn as nn
import torch.nn.functional as F# define the CNN architecture
class Net(nn.Module):
    def __init__(self, hidden_dim, dropout =0.3):
        
        super(Net, self).__init__()
        # Number of features
        self.fc1 = nn.Linear(427, hidden_dim)
        
        self.fc2 = nn.Linear(hidden_dim, hidden_dim*2)
        
        self.fc3 = nn.Linear(hidden_dim*2, hidden_dim)
        
        self.fc4 = nn.Linear(hidden_dim, 32)
        
        self.fc5 = nn.Linear(32, 3)
        
        self.dropout = nn.Dropout(dropout)
        
    def forward(self, x):
        
        out = self.dropout(F.relu(self.fc1(x)))
        
        out = self.dropout(F.relu(self.fc2(out)))
        
        out = self.dropout(F.relu(self.fc3(out)))
        
        out = self.dropout(F.relu(self.fc4(out)))
        
        out = self.fc5(out)
        
        return out

After creating the model and required transformations, we will create our training loop

We are going to train our model in AWS Sagemaker. It is completely an optional step. The model can be trained locally, and the model output file can be used for predictions. If you train it in the cloud, below code can be used for Training.

You also need an AWS account with the Sagemaker setup. If you need more info or help, please check my previous article Train a GAN and generate faces using AWS Sagemaker | PyTorch setup section.

Once you have all the required access, you can start fitting the model as shown below. Below command will package all the necessary code with data and create an EC2 server with required containers and train the model.

from sagemaker.pytorch import PyTorch#Check the status of dataloader
estimator = PyTorch(entry_point="train.py",
                    source_dir="train",
                    role=role,
                    framework_version='1.0.0',
                    train_instance_count=1,
                    train_instance_type='ml.p2.xlarge',
                    hyperparameters={
                        'epochs': 2,
                        'hidden_dim': 32,
                    },)

Once you train the model, all the corresponding files will be in your S3 bucket. If you train your model locally, make sure you have the files in the corresponding S3 bucket location.

24x5 Stock Trading Agent to predict stock prices with Deep Learning with deployment

AWS Cloud to host code and get predictions

As our next setup, we will deploy the model in AWS Sagemaker. When deploying a PyTorch model in SageMaker, you are expected to provide four functions that the SageMaker inference container will use.

model_fn
input_fn
output_fn
predict_fn

Below is the code to load the model and prepare the input data

Some points to be noted in the above code:

  1. The model and scaler object need to be in an S3 bucket.
  2. We have fetched data for many days or hours. It is required for LSTM type networks.
  3. Input content is the ticker symbol. We can tune the code for multiple symbols.

In below code section, we will create the output and predict function

Some points to be noted in the above code

  1. We have three classes buy, sell or hold. Prediction needs to be one of these three.
  2. We need to focus on what is predicted and returned.
  3. Trade only if there are enough funds (or limited funds) and in limited quantity.

Deployment is similar to training the model

from sagemaker.predictor import RealTimePredictor
from sagemaker.pytorch import PyTorchModelclass StringPredictor(RealTimePredictor):
    def __init__(self, endpoint_name, sagemaker_session):
        super(StringPredictor, self).__init__(endpoint_name, sagemaker_session, content_type='text/plain')model = PyTorchModel(model_data=estimator.model_data,
                     role = role,
                     framework_version='1.0.0',
                     entry_point='predict.py',
                     source_dir='../serve',
                     predictor_cls=StringPredictor,)# Deploy the model in cloud serverpredictor = model.deploy(initial_instance_count=1, instance_type='ml.m5.large')

If you want to test the model, you can execute the below code. This says that the workflow is working, and you can output your predictions and trade stocks on the issued ticker.

24x5 Stock Trading Agent to predict stock prices with Deep Learning with deployment

Prediction endpoint name

You can also get the endpoint name from the above code in the screenshot.

Create a Lambda function and API

Here will complete the pipeline be creating the Lamda function and API.

Create a Lambda function

Create a lambda function in the AWS lambda service. Remember to update the Endpoint name from the above screenshot.

API Gateway

From AWS API gateway services, create a Rest API. Then give a name and create the API.

24x5 Stock Trading Agent to predict stock prices with Deep Learning with deployment

24x5 Stock Trading Agent to predict stock prices with Deep Learning with deployment

Create a post and deploy from Actions dropdown. Once you create it, the API is ready. You can use it in any UI if required.

Finally, we have our Rest endpoint, where we can create post requests. The endpoint can be tested with Postman or any other tool. If you don’t need an endpoint, you can schedule the lambda function by following this link .

24x5 Stock Trading Agent to predict stock prices with Deep Learning with deployment

You can see that the stock is bought and sold in the Alpaca portal. Predictions are predicted from our model, and live data is fed into the model.

If there is enough interest in the article, I will write a followup article on adding sentiment analysis about that particular stock in real-time with backtesting and add other model architectures.

以上就是本文的全部内容,希望本文的内容对大家的学习或者工作能带来一定的帮助,也希望大家多多支持 码农网

查看所有标签

猜你喜欢:

本站部分资源来源于网络,本站转载出于传递更多信息之目的,版权归原作者或者来源机构所有,如转载稿涉及版权问题,请联系我们

嵌入式Linux应用程序开发详解

嵌入式Linux应用程序开发详解

孙琼 / 人民邮电出版社 / 2006-7 / 46.00元

《嵌入式Linux应用程序开发详解》主要分为3个部分,包括Linux基础、搭建嵌入式Linux环境和嵌入式Linux的应用开发。Linux基础部分从Linux的安装过程、基本操作命令讲起,为Linux初学者能快速入门提供了保证。接着系统地讲解了嵌入式Linux的环境搭建,以及嵌入式Linux的I/O与文件系统的开发、进程控制开发、进程间通信开发、网络应用开发、基于中断的开发、设备驱动程序的开发以及......一起来看看 《嵌入式Linux应用程序开发详解》 这本书的介绍吧!

随机密码生成器
随机密码生成器

多种字符组合密码

XML、JSON 在线转换
XML、JSON 在线转换

在线XML、JSON转换工具

HEX HSV 转换工具
HEX HSV 转换工具

HEX HSV 互换工具