Automated trading using deep learning and machine learning, building a trading agent using reinforcement learning Implementing a trading agent that learns autonomously using reinforcement learning techniques.

1. Introduction

The cryptocurrency market, such as Bitcoin, is highly volatile, and various technologies are being researched to automate trading. Deep Learning and Machine Learning techniques are effective in building such automated trading systems. This post explains how to build a self-learning trading agent using Reinforcement Learning techniques.

2. Basics of Machine Learning and Deep Learning

Machine Learning is a methodology for learning patterns from data and creating predictive models. Deep Learning is a subfield of Machine Learning that uses artificial neural networks to learn the structure of complex data. Their advantage is the ability to process large amounts of data.

2.1. Understanding Reinforcement Learning

Reinforcement Learning is a method where an agent learns the optimal actions through interaction with the environment. The agent selects specific actions from a given state and receives rewards as a result. Through this reward, the agent improves its actions.

3. Building a Trading Agent Based on Reinforcement Learning

3.1. Configuring the Environment

Configuring the environment for the trading agent is very important. To this end, we define the market environment based on OHLC (Open, High, Low, Close) data.

3.2. Installing OpenAI Gym

You can use OpenAI’s Gym library to create a reinforcement learning environment. Installation can be done via the following command.

pip install gym

3.3. Implementing the Trading Environment

Below is a code that implements a simple trading environment.


import gym
from gym import spaces
import numpy as np

class CryptoTradingEnv(gym.Env):
    def __init__(self, data):
        super(CryptoTradingEnv, self).__init__()
        self.data = data
        self.current_step = 0
        self.action_space = spaces.Discrete(3)  # 0: Hold, 1: Buy, 2: Sell
        self.observation_space = spaces.Box(low=0, high=np.inf, shape=(len(data[0]),), dtype=np.float32)

    def reset(self):
        self.current_step = 0
        return self.data[self.current_step]

    def step(self, action):
        self.current_step += 1
        if self.current_step >= len(self.data):
            self.current_step = len(self.data) - 1
        
        prev_state = self.data[self.current_step - 1]
        current_state = self.data[self.current_step]

        reward = 0
        if action == 1:  # Buy
            reward = current_state[3] - prev_state[3]  # Close price
        elif action == 2:  # Sell
            reward = prev_state[3] - current_state[3]

        done = self.current_step == len(self.data) - 1
        return current_state, reward, done, {}
    

3.4. Building the Deep Learning Model

Now we implement a deep learning model to train the reinforcement learning agent. Here, we use a simple Multi-layer Perceptron (MLP).


import tensorflow as tf
from tensorflow import keras
from tensorflow.keras import layers

def create_model(input_shape):
    model = keras.Sequential()
    model.add(layers.Dense(24, activation='relu', input_shape=input_shape))
    model.add(layers.Dense(24, activation='relu'))
    model.add(layers.Dense(3, activation='linear'))  # 3 actions
    model.compile(optimizer='adam', loss='mse')
    return model
    

3.5. Training the Agent

The agent learns its policy through multiple episodes. Here, we apply a simple Q-learning algorithm.


import random

class DQNAgent:
    def __init__(self, state_size):
        self.state_size = state_size
        self.memory = []
        self.gamma = 0.95  # discount rate
        self.epsilon = 1.0  # exploration rate
        self.epsilon_min = 0.01
        self.epsilon_decay = 0.995
        self.model = create_model((state_size,))

    def remember(self, state, action, reward, next_state, done):
        self.memory.append((state, action, reward, next_state, done))

    def act(self, state):
        if np.random.rand() <= self.epsilon:
            return random.randrange(3)  # exploration
        q_values = self.model.predict(state)
        return np.argmax(q_values[0])  # exploitation

    def replay(self, batch_size):
        minibatch = random.sample(self.memory, batch_size)
        for state, action, reward, next_state, done in minibatch:
            target = reward
            if not done:
                target += self.gamma * np.amax(self.model.predict(next_state)[0])
            target_f = self.model.predict(state)
            target_f[0][action] = target
            self.model.fit(state, target_f, epochs=1, verbose=0)
        
        if self.epsilon > self.epsilon_min:
            self.epsilon *= self.epsilon_decay
    

4. Conclusion

This post explained the automatic trading system for Bitcoin using reinforcement learning techniques. We built a simple trading environment and a deep learning model, and covered the approach of learning using Q-learning. More data and hyperparameter tuning are needed to predict actual Bitcoin prices and establish trading strategies. Lastly, exchange API integration will be necessary for real trading.

5. References

Automated trading using deep learning and machine learning, combining reinforcement learning and momentum strategies to improve the performance of momentum-based trading strategies through reinforcement learning.

1. Introduction

In recent years, the popularity of cryptocurrencies like Bitcoin has surged. Additionally,
machine learning and deep learning techniques have gained attention in the financial sector,
leading many investors to utilize these technologies to develop automated trading systems.
This article will explore methods to enhance the performance of momentum-based trading strategies
through reinforcement learning.

2. Basic Concepts

2.1. Machine Learning and Deep Learning

Machine learning is the field that develops algorithms to learn patterns and make predictions from
data. In contrast, deep learning is a subset of machine learning that utilizes artificial neural
networks to learn complex patterns. These two technologies serve as powerful tools for data
analysis and prediction.

2.2. Reinforcement Learning

Reinforcement learning is a method where an agent learns to maximize rewards by interacting with
the environment. In this process, the agent learns the impact of its actions on the outcomes.
This approach is suitable for automated trading systems, as it can harness market volatility to
pursue profits.

2.3. Momentum Strategy

The momentum strategy is an investment technique that predicts future prices based on past price trends.
Generally, it involves buying assets believing that the uptrend will continue and selling them
believing that the downtrend will persist. This strategy includes purchasing assets that are
rising in price over a certain period.

3. Combining Reinforcement Learning and Momentum Strategy

3.1. System Design

When designing an automated trading system, the first step is to define the environment.
This environment consists of price data and trading information, and the agent will make trading
decisions within this environment. The agent’s ultimate goal is to achieve the maximum reward.

3.2. Data Collection

Bitcoin price data can be collected from various sources.
Here, we will collect price data through a simple API and use it for training the reinforcement
learning model. The data may consist of historical prices, trading volume, etc.

3.3. Defining States and Actions

The agent selects actions based on the current state.
The state is defined using price data along with technical indicators (moving average, RSI, etc.),
and actions can be set as buying, selling, or holding.

3.4. Designing the Reward Function

The reward function serves as a criterion to assess how successful the agent’s actions are.
Typically, it is designed to reward the agent when a profit is made after buying, and impose
a penalty when a loss occurs. The reward can be based on trading profits and losses.

4. Example Code

Below is a simple example code for automated trading of Bitcoin using reinforcement learning.
This code structures the environment using OpenAI’s Gym and demonstrates how to train the agent using
the deep learning library TensorFlow.

        
        import numpy as np
        import pandas as pd
        import gym
        from gym import spaces
        from tensorflow.keras import Sequential
        from tensorflow.keras.layers import Dense
        from tensorflow.keras.optimizers import Adam

        class BitcoinEnv(gym.Env):
            def __init__(self, data):
                super(BitcoinEnv, self).__init__()
                self.data = data
                self.action_space = spaces.Discrete(3)  # 0: Sell, 1: Buy, 2: Hold
                self.observation_space = spaces.Box(low=0, high=1, shape=(data.shape[1],), dtype=np.float32)
                self.current_step = 0
                self.balance = 1000  # Initial capital
                self.position = 0  # Current holdings

            def reset(self):
                self.current_step = 0
                self.balance = 1000
                self.position = 0
                return self.data[self.current_step]

            def step(self, action):
                current_price = self.data[self.current_step]['close']
                reward = 0

                if action == 1:  # Buy
                    self.position = self.balance / current_price
                    self.balance = 0
                elif action == 0:  # Sell
                    if self.position > 0:
                        self.balance = self.position * current_price
                        reward = self.balance - 1000  # Profit
                        self.position = 0

                self.current_step += 1
                done = self.current_step >= len(self.data) - 1
                next_state = self.data[self.current_step]
                return next_state, reward, done, {}

        # Define a simple neural network model.
        def build_model(input_shape):
            model = Sequential()
            model.add(Dense(24, input_shape=input_shape, activation='relu'))
            model.add(Dense(24, activation='relu'))
            model.add(Dense(3, activation='linear'))  # 3 actions
            model.compile(optimizer=Adam(lr=0.001), loss='mse')
            return model

        # Main execution code
        if __name__ == "__main__":
            # Load data
            data = pd.read_csv('bitcoin_price.csv')  # Bitcoin price data
            env = BitcoinEnv(data)
            model = build_model((data.shape[1],))

            # Agent training
            for episode in range(1000):
                state = env.reset()
                done = False

                while not done:
                    action = np.argmax(model.predict(state.reshape(1, -1)))
                    next_state, reward, done, _ = env.step(action)
                    model.fit(state.reshape(1, -1), reward, verbose=0)  # Simple training
                    state = next_state
        
    

5. Result Analysis

After running the code, various metrics can be used to analyze how efficiently the agent traded Bitcoin.
For example, the final return, maximum drawdown, and Sharpe ratio can be calculated to evaluate the
performance of the strategy.

6. Conclusion

This course introduced methods to improve momentum-based trading strategies through reinforcement learning.
It demonstrated how machine learning and deep learning technologies can be utilized in automated trading
in financial markets, and provided hints on future research directions.
This field still has great potential for development, and more innovative automated trading systems can be
developed through various techniques.

7. References

  • 1. Sutton, R. S., & Barto, A. G. (2018). Reinforcement Learning: An Introduction.
  • 2. Goodfellow, I., Yoshua Bengio, & Aaron Courville. (2016). Deep Learning.
  • 3. Bitcoin historical data source: CoinGecko.

Automated trading using deep learning and machine learning, price prediction based on Gaussian Process Regression (GPR) Applying Gaussian Process Regression to predict the price movements of Bitcoin.

To build an automated trading system for cryptocurrencies like Bitcoin, an effective price prediction model is essential. This article will detail how to predict Bitcoin’s price fluctuations using Gaussian Process Regression (GPR), one of the machine learning techniques.

1. Overview of Machine Learning and Deep Learning

Machine learning is a field of artificial intelligence (AI) that enables predictions on new data by learning patterns from data. Deep learning is a subset of machine learning that uses artificial neural networks to learn the features of complex data independently.

2. What is Gaussian Process Regression (GPR)?

Gaussian Process Regression (GPR) is a form of nonparametric Bayesian statistical model, particularly effective for predicting continuous data. GPR creates a probabilistic model for the given data, naturally incorporating uncertainty. This allows for estimating the confidence level of predictions alongside predicted values.

2.1 Mathematical Background of GPR

GPR is based on Gaussian distribution and learns the functional relationship between input data and output data. For the given training dataset (X, y), GPR uses the following covariance function for predictions:

K(X, X') = σ² * exp(-||X - X'||² / (2 * l²))

Here, K is the kernel function, σ is the standard deviation of noise, and l is the length scale. This kernel function determines the similarity between data points.

3. Collecting Bitcoin Price Data

To build a Bitcoin price prediction model, historical Bitcoin price data is required. We will use the pandas library and the yfinance module in Python to collect data.

import pandas as pd
import yfinance as yf

# Download Bitcoin data
btc_data = yf.download('BTC-USD', start='2020-01-01', end='2023-01-01')
btc_data = btc_data[['Close']]
btc_data = btc_data.rename(columns={'Close': 'price'})
btc_data = btc_data.reset_index()
btc_data['Date'] = pd.to_datetime(btc_data['Date'])
btc_data.sort_values('Date', inplace=True)
print(btc_data.head())

4. Data Preprocessing

The collected data must be preprocessed to fit the GPR model. In particular, for time series data, trends and seasonality may need to be removed.

btc_data['returns'] = btc_data['price'].pct_change()
btc_data = btc_data.dropna()

# Reset index
btc_data.reset_index(drop=True, inplace=True)
print(btc_data.head())

5. Building the Gaussian Process Regression Model

To build the model, we will use the GaussianProcessRegressor class from the scikit-learn library. This allows us to predict Bitcoin prices.

from sklearn.gaussian_process import GaussianProcessRegressor
from sklearn.gaussian_process.kernels import RBF, ConstantKernel as C

# Define kernel
kernel = C(1.0, (1e-3, 1e3)) * RBF(1.0, (1e-2, 1e2))

# Initialize model
gpr = GaussianProcessRegressor(kernel=kernel, n_restarts_optimizer=10)

# Training data
X_train = btc_data.index.values.reshape(-1, 1)
y_train = btc_data['price'].values

# Fit model
gpr.fit(X_train, y_train)

6. Price Prediction

Let’s use the trained GPR model to predict future prices. We will decide on a date for prediction and create an index to perform the prediction.

import numpy as np

# Number of days to predict
n_days = 30
X_test = np.arange(len(btc_data), len(btc_data) + n_days).reshape(-1, 1)

# Prediction
y_pred, sigma = gpr.predict(X_test, return_std=True)

# Visualize results
import matplotlib.pyplot as plt

plt.figure(figsize=(12, 6))
plt.plot(btc_data['Date'], btc_data['price'], 'r.', markersize=10, label='Observed Data')
plt.plot(btc_data['Date'].iloc[-1] + pd.to_timedelta(np.arange(1, n_days + 1), unit='D'), y_pred, 'b-', label='Predicted Price')
plt.fill_between(btc_data['Date'].iloc[-1] + pd.to_timedelta(np.arange(1, n_days + 1), unit='D'),
                 y_pred - 2 * sigma, y_pred + 2 * sigma, color='gray', alpha=0.2, label='Confidence Interval')
plt.title('Bitcoin Price Prediction using Gaussian Process Regression')
plt.xlabel('Date')
plt.ylabel('Price in USD')
plt.legend()
plt.show()

7. Performance Evaluation

To evaluate the model’s performance, we can use the Root Mean Squared Error (RMSE) and R² Score. This can help gauge the accuracy of the predictions.

from sklearn.metrics import mean_squared_error, r2_score

# Calculate RMSE
y_train_pred = gpr.predict(X_train)
rmse = np.sqrt(mean_squared_error(y_train, y_train_pred))
r2 = r2_score(y_train, y_train_pred)

print(f"RMSE: {rmse:.2f}, R² Score: {r2:.2f}")

8. Building a Real-time Automated Trading System

Finally, automated trading can be implemented based on the predicted prices. This should include logic to generate trading signals (buy/sell) and interface with exchanges through APIs for actual trading.

def generate_signals(predicted_prices):
    buy_signals = []
    sell_signals = []
    for i in range(1, len(predicted_prices)):
        if predicted_prices[i] > predicted_prices[i - 1]:
            buy_signals.append(predicted_prices[i])
            sell_signals.append(np.nan)
        elif predicted_prices[i] < predicted_prices[i - 1]:
            sell_signals.append(predicted_prices[i])
            buy_signals.append(np.nan)
        else:
            buy_signals.append(np.nan)
            sell_signals.append(np.nan)
    return buy_signals, sell_signals

buy_signals, sell_signals = generate_signals(y_pred)

plt.figure(figsize=(12, 6))
plt.plot(btc_data['Date'], btc_data['price'], label='Actual Price')
plt.plot(btc_data['Date'].iloc[-1] + pd.to_timedelta(np.arange(1, n_days + 1), unit='D'), y_pred, label='Predicted Price', color='orange')
plt.plot(btc_data['Date'].iloc[-1] + pd.to_timedelta(np.arange(1, n_days + 1), unit='D'), buy_signals, marker='^', color='g', label='Buy Signal', markersize=10)
plt.plot(btc_data['Date'].iloc[-1] + pd.to_timedelta(np.arange(1, n_days + 1), unit='D'), sell_signals, marker='v', color='r', label='Sell Signal', markersize=10)
plt.title('Buy/Sell Signals based on Predictions')
plt.xlabel('Date')
plt.ylabel('Price in USD')
plt.legend()
plt.show()

9. Conclusion

In this tutorial, we explored how to build a Bitcoin price prediction model using Gaussian Process Regression. GPR has the advantage of effectively reflecting the uncertainty of price predictions and can be applied to automated trading systems.

In the future, adding more features and testing other machine learning algorithms could be beneficial to improve this system. Additionally, integrating real-time data could help implement a more effective automated trading system.

Finally, remember that trading stocks or cryptocurrencies always involves risks. It is important to operate an automated trading system after sufficient research and testing.

Automatic trading using deep learning and machine learning, configuring reinforcement learning environments, and training agents. Creating a Bitcoin trading environment using OpenAI Gym and the reinforcement learning training process.

In today’s financial markets, algorithmic trading and automated trading strategies have become major topics. Especially in the cryptocurrency market, such as Bitcoin, quick decision-making and execution are essential. This article will explore how to perform automated trading of Bitcoin using deep learning and machine learning techniques, and explain how to set up a reinforcement learning environment based on OpenAI Gym and train agents.

1. The Need for Automated Bitcoin Trading

Automated Bitcoin trading aims for traders to make immediate trading decisions based on market analysis. By excluding human emotions and analyzing data through algorithms, better trading decisions can be made. Recently, machine learning and deep learning techniques have been applied in this field, leading to more sophisticated predictive models.

2. Understanding Reinforcement Learning (Deep Reinforcement Learning)

Reinforcement learning is a machine learning technique where an agent learns optimal decision-making by interacting with the environment. The agent receives reward signals and adjusts its actions, learning the optimal policy. In Bitcoin trading, actions such as buy, sell, or wait are chosen based on price fluctuations or other market indicators.

3. Setting Up a Bitcoin Trading Environment Using OpenAI Gym

OpenAI Gym is a toolkit that provides various reinforcement learning environments. Through this, a Bitcoin trading environment can be set up, allowing agents to learn within this environment. The essential elements needed to create a Bitcoin trading environment using OpenAI Gym can be summarized as follows.

  1. Environment Setup: Collect Bitcoin price data to configure the Gym environment. This data defines the agent’s state and designs the reward structure.
  2. Action Definition: Define actions such as buy, sell, and wait so that the agent can choose from them in each state.
  3. Reward Structure Design: Define the rewards obtained based on the agent’s actions. For example, provide positive rewards for profits and negative rewards for losses.

3.1. Example Code: Bitcoin Trading Environment

    
    import numpy as np
    import gym
    from gym import spaces

    class BitcoinTradingEnv(gym.Env):
        def __init__(self, data):
            super(BitcoinTradingEnv, self).__init__()
            self.data = data
            self.current_step = 0
            
            # Define action space: 0 - wait, 1 - buy, 2 - sell
            self.action_space = spaces.Discrete(3)
            
            # Define observation space: current balance, holding amount, price
            self.observation_space = spaces.Box(low=0, high=np.inf, shape=(3,), dtype=np.float32)

        def reset(self):
            self.current_step = 0
            self.balance = 1000  # Initial balance
            self.holding = 0      # Holding Bitcoin
            return self._get_observation()

        def _get_observation(self):
            price = self.data[self.current_step]
            return np.array([self.balance, self.holding, price])

        def step(self, action):
            current_price = self.data[self.current_step]
            reward = 0

            if action == 1:  # Buy
                if self.balance >= current_price:
                    self.holding += 1
                    self.balance -= current_price
                    reward = -1  # Cost: buy
            elif action == 2:  # Sell
                if self.holding > 0:
                    self.holding -= 1
                    self.balance += current_price
                    reward = 1  # Profit: sell

            self.current_step += 1
            done = self.current_step >= len(self.data)
            return self._get_observation(), reward, done, {}

    # Example usage
    data = np.random.rand(100) * 100  # Simulated price data
    env = BitcoinTradingEnv(data)
    
    

4. Training Agents Using Deep Learning Models

To train a reinforcement learning agent, deep learning models can be applied to learn policies or values. Here, the method using the DQN (Deep Q-Network) algorithm will be explained. DQN integrates the Q-learning algorithm with a deep learning model, taking the state as input and outputting Q values.

4.1. Example Code: DQN Algorithm

    
    import numpy as np
    import tensorflow as tf
    from collections import deque

    class DQNAgent:
        def __init__(self, action_size):
            self.action_size = action_size
            self.state_size = 3
            self.memory = deque(maxlen=2000)
            self.gamma = 0.95  # Discount rate
            self.epsilon = 1.0  # Exploration rate
            self.epsilon_min = 0.01
            self.epsilon_decay = 0.995
            self.model = self._build_model()

        def _build_model(self):
            model = tf.keras.Sequential()
            model.add(tf.keras.layers.Dense(24, input_dim=self.state_size, activation='relu'))
            model.add(tf.keras.layers.Dense(24, activation='relu'))
            model.add(tf.keras.layers.Dense(self.action_size, activation='linear'))
            model.compile(loss='mse', optimizer=tf.keras.optimizers.Adam(lr=0.001))
            return model

        def remember(self, state, action, reward, next_state, done):
            self.memory.append((state, action, reward, next_state, done))

        def act(self, state):
            if np.random.rand() <= self.epsilon:
                return np.random.choice(self.action_size)
            act_values = self.model.predict(state)
            return np.argmax(act_values[0])

        def replay(self, batch_size):
            minibatch = np.random.choice(len(self.memory), batch_size)
            for index in minibatch:
                state, action, reward, next_state, done = self.memory[index]
                target = reward
                if not done:
                    target += self.gamma * np.amax(self.model.predict(next_state)[0])
                target_f = self.model.predict(state)
                target_f[0][action] = target
                self.model.fit(state, target_f, epochs=1, verbose=0)
            if self.epsilon > self.epsilon_min:
                self.epsilon *= self.epsilon_decay

    # Example usage
    agent = DQNAgent(action_size=3)
    
    

4.2. Agent Learning Process

The agent learns through multiple episodes. In each episode, the environment is reset, and the state, reward, and next state are obtained based on the agent’s actions. This information is remembered, and the model is learned by sampling the specified batch size.

Below is a basic structure for training the agent and evaluating performance:

    
    episodes = 1000
    batch_size = 32

    for e in range(episodes):
        state = env.reset()
        state = np.reshape(state, [1, agent.state_size])
        for time in range(500):
            action = agent.act(state)
            next_state, reward, done, _ = env.step(action)
            next_state = np.reshape(next_state, [1, agent.state_size])
            agent.remember(state, action, reward, next_state, done)
            state = next_state
            if done:
                print(f'Episode: {e}/{episodes}, Score: {time}, epsilon: {agent.epsilon:.2}')
                break
            if len(agent.memory) > batch_size:
                agent.replay(batch_size)
    
    

5. Conclusion

This tutorial explained how to build an automated trading system for Bitcoin using deep learning and machine learning, and how to set up a reinforcement learning environment using OpenAI Gym and train agents. Applying reinforcement learning in Bitcoin trading is still a field with much research, and various strategies and approaches can be experimented with to achieve success in the real world.

We look forward to how your systems can evolve in the future, and hope you make smarter investment decisions through machine learning and deep learning technologies.

Automated Trading Using Deep Learning and Machine Learning, Trading Prediction Using XGBoost How to Generate High-Performance Trading Signals Using XGBoost.

Recently, automated trading systems through artificial intelligence (AI), deep learning, and machine learning have rapidly developed in financial markets. These technologies are powerful tools that can learn patterns from data and make trading decisions based on that learning. In this blog post, we will take an in-depth look at how to automatically trade cryptocurrencies like Bitcoin using XGBoost (Extreme Gradient Boosting).

What is Automated Trading?

An automated trading system is software that conducts trades through pre-set algorithms. Emotional decisions by humans are excluded, and decisions are made based on data. Such automated trading predicts market trends through high-frequency trading, pattern recognition, and technical analysis like Bollinger Bands.

What is XGBoost?

XGBoost is an extension of the Gradient Boosting algorithm, and it is a powerful predictive model often used in machine learning competitions. The reasons for its superior performance are as follows:

  • Accuracy: It creates better models through regularization with the loss function.
  • Scalability: It is efficient at handling large datasets.
  • Parallel Processing: It utilizes multiple CPU cores to enhance learning speed.

Generating Trading Signals Using XGBoost

The goal of automated trading is to generate buy or sell signals. XGBoost can learn from historical data to predict future prices. Here is the signal generation process using XGBoost.

Step 1: Data Collection

First, we need to collect Bitcoin price data. Here, we will show an example of fetching data via the Binance API.


import numpy as np
import pandas as pd
import requests

def fetch_data(symbol, interval, start, end):
    url = f'https://api.binance.com/api/v3/klines?symbol={symbol}&interval={interval}&startTime={start}&endTime={end}'
    response = requests.get(url)
    data = response.json()
    df = pd.DataFrame(data, columns=['open_time', 'open', 'high', 'low', 'close', 'volume', 'close_time', 'quote_asset_volume', 'number_of_trades', 'taker_buy_base_asset_volume', 'taker_buy_quote_asset_volume', 'ignore'])
    df['close'] = df['close'].astype(float)
    return df

# Example: Fetching daily data for BTCUSDT.
data = fetch_data('BTCUSDT', '1d', '1609459200000', '1640995200000')  # From January 1, 2021, to January 1, 2022.

Step 2: Data Preprocessing

Extract the necessary features from the collected data. For example, technical indicators such as moving averages, RSI, and MACD can be calculated.


def compute_features(df):
    df['MA5'] = df['close'].rolling(window=5).mean()
    df['MA20'] = df['close'].rolling(window=20).mean()
    df['RSI'] = compute_rsi(df['close'])
    df['MACD'] = compute_macd(df['close'])
    return df.dropna()

def compute_rsi(series, period=14):
    delta = series.diff()
    gain = (delta.where(delta > 0, 0)).rolling(window=period).mean()
    loss = (-delta.where(delta < 0, 0)).rolling(window=period).mean()
    rs = gain / loss
    rsi = 100 - (100 / (1 + rs))
    return rsi

def compute_macd(series):
    exp1 = series.ewm(span=12, adjust=False).mean()
    exp2 = series.ewm(span=26, adjust=False).mean()
    return exp1 - exp2

data = compute_features(data)

Step 3: Splitting Training and Testing Data

To train the model, split the data into training and testing sets. Typically, 70% to 80% of the data is used for training.


from sklearn.model_selection import train_test_split

X = data[['MA5', 'MA20', 'RSI', 'MACD']].values
y = np.where(data['close'].shift(-1) > data['close'], 1, 0)[:-1]  # If the price rises the next day

X_train, X_test, y_train, y_test = train_test_split(X[:-1], y, test_size=0.2, random_state=42)

Step 4: Training the XGBoost Model

Now we will train the XGBoost model. XGBoost creates high-performance predictors.


from xgboost import XGBClassifier

model = XGBClassifier()
model.fit(X_train, y_train)

Step 5: Generating Trading Signals

Use the trained model to generate trading signals. Based on the prediction results, we can assign buy and sell signals.


predictions = model.predict(X_test)
predictions_proba = model.predict_proba(X_test)

buy_signals = np.where(predictions == 1, 1, 0)  # Buy signal
sell_signals = np.where(predictions == 0, -1, 0)  # Sell signal

signals = buy_signals + sell_signals

Step 6: Strategy Validation

Compare the generated trading signals with actual price data to validate the strategy’s performance. This process is called backtesting and is an important step in evaluating the model’s validity.


def backtest(signals, prices):
    initial_capital = 10000
    shares = 0
    capital = initial_capital

    for i in range(len(signals)):
        if signals[i] == 1:  # Buy signal
            shares += capital // prices[i]
            capital -= (capital // prices[i]) * prices[i]
        elif signals[i] == -1:  # Sell signal
            capital += shares * prices[i]
            shares = 0

    return capital + (shares * prices[-1])

strategy_return = backtest(signals, data['close'].values[len(X_train):])
print('Strategy Return:', strategy_return)

Conclusion

Automated trading systems utilizing deep learning and machine learning technologies can enable data-driven decision-making, thereby maximizing investors’ profitability. Among these, XGBoost shows outstanding performance and is effective in generating trading signals for highly volatile assets like Bitcoin.

Based on this material, we encourage you to improve your algorithm and apply it to various assets. Continuous learning and experimentation are necessary to succeed in the world of automated trading.