layout | title | permalink |
---|---|---|
page |
Overview |
/ |
The submission deadline has been extended to Nov 7th, 2024, 23:59 EST!
Task 1 and 2 starter kits are released here!
Financial Reinforcement Learning (FinRL) is an interdisciplinary field that applies reinforcement learning to perform financial tasks. FinRL’s ability to adapt to changing market conditions makes it a powerful tool for developing strategies in various tasks, such as portfolio management, algorithmic trading, and option pricing.
The FinRL contest is a competition that explores and evaluates the capability of machine learning methods in finance. FinRL Contest 2024 introduces two tasks designed to address key challenges in FinRL. It features with:
- Ensemble Learning: Tackling the challenge of policy instability in FinRL, ensemble learning can reduce the variance and bias associated with individual agents, leading to more reliable performance in volatile market conditions. To overcome the sampling bottleneck and accelerate the training of component agents, we also provide a vectorized environment that supports massively parallel simulations.
- Reinforcement Learning from Market Feedback (RLMF): Large language models (LLMs) have been used in financial tasks, such as sentiment analysis and generating novel trading signals. However, there is a gap between general-purpose LLMs and financial markets, since general-purpose LLMs trained on Internet data may not capture the intrinsic dynamics of financial markets. To align LLMs with financial markets, we propose to adapt LLMs using Reinforcement Learning from Market Feedback, as a counterpart of Reinforcement Learning from Human Feedback (RLHF). RLMF utilizes feedback from the financial market as reward signals, enabling LLMs to learn from financial market behaviors.
We design two tasks to reflect these advancements: (1) Cryptocurrency trading with ensemble methods, and (2) LLM-engineered signals with RLMF. We welcome students, researchers, and engineers who are passionate about finance and machine learning. And we encourage the development of ensemble strategies, novel signals, and innovative algorithms that can adapt to changing market conditions and generate superior returns for investors.
We have more than 30 market data sources to deal with different financial tasks. We hold the data APIs and sample market environments in an open-source repository, FinRL-Meta, as shown in Figure 1. In addition, in FinGPT, as shown in Figure 2, we have a variety of financial data sources that assures comprehensive market coverage. Contestants are welcome to explore and use in the FinRL Contest.
With a deep reinforcement learning approach, market datasets are processed into gym-style environments. The market environment provided to participants is designed to enhance both the realism and efficiency of the simulation process.
Table 1 lists the state spaces, action spaces, and reward functions of different FinRL applications. A state shows how an agent perceives a market situation. Facing a state, the agent can take an action from the action set, which may vary according to the financial tasks. Reward is an incentive mechanism for an agent to learn a better policy. Contestants will specify the state space, action space, and reward functions in the environment for both tasks.
Figure 3 shows the provided vectorized environment to support massively parallel simulataions. It manages multiple parallel sub-environments, each simulating different market scenarios and incorporating realistic market constraints.
- Team Registration Begin: October 6th, 2024
- Starter-Kit Release: October 15th, 2024
- Submission Open: October 20th, 2024
- Solution Submission Deadline:
November 3rd, 2024November 7th, 2024 - Report Submission Deadline:
November 7th, 2024November 8th, 2024 - Winner Notification: November 12th, 2024
- Winner Announcement: November 15th, 2024
(All deadlines are at 11:59pm EST on the specified date.)
Winners will be invited to attend the ACM ICAIF 2024 conference and have the opportunity to present their work at the conference.
Each team can choose to participate in one or both tasks. The prizes will be awarded for each task.
We provide some tutorials for participants to learn FinRL.
The starter kits are released here
This task aims to develop robust and effective trading agents for cryptocurrencies using ensemble methods. Participants are expected to explore innovative ensemble methods for single cryptocurrency trading. They are also encouraged to take advantage of the power of massively parallel simulations by utilizing the provided vectorized environments.
A dataset containing second-level Limit Order Book (LOB) data for Bitcoin is provided. Contestants are free to apply various techniques to the data, design component models, and use innovative methods to increase the diversity of component models in the ensemble.
The initial cash should be $1 million.
The contestants are required to:
- Specify the state space, action space, and reward functions in the environment.
- Ensure that the final ensemble model should be able to interact with the provided trading environment.
The performance of the model will be assessed by the following metrics:
- Cumulative return. It is the total return generated by the trading strategy over a trading period.
- Sharpe ratio. It takes into account both the returns of the portfolio and the level of risk.
- Max drawdown. It is the portfolio’s largest percentage drop from a peak to a trough in a certain time period, which provides a measure of downside risk.
Participants need to submit a well-organized repository containing all scripts, models, and any custom libraries used to implement the solution. Each team should also submit a 1-2 page report with the ACM sigconf template through Open Review. The title should start with “FinRL Contest 2024 Task I.”
This task aims to develop LLMs that can generate and engineer effictive signals from news by using Reinforcement Learning from Market Feedback (RLMF). By incorporating market feedback in the fine-tuning process, LLMs can learn from and adapt to financial market behaviors.
In this task, the LLM will be used to generate one type of signal (e.g., a sentiment score) based on the content of news. Contestants will develop models that leverage RLMF to adjust the signals based on the market feedback. Contestants are expected to explore useful market feedback from market data and innovative reward models to fine-tune their LLMs.
An OHLCV dataset and a corresponding news dataset for a list of stocks are provided. Contestants are free to use external datasets to deveop RLMF methods and fine-tune the LLMs.
- Contestants should use the LLM that can be loaded with huggingface and can do inference on 20GB GPU.
- Contestants can determine what signal to generate and engineer.
- Contestants should provide the prompt for their LLM. For example, "What is the sentiment score of {stock ticker} after the release of this news: {news}. Give and only return the score in the range of [-1, 1]. Answer: "
To assess the effectiveness of the signal engineered by the LLMs, we will apply the signal to a simple and practical trading strategy:
- Buy the top 3 stocks with the highest signals on the day of the news release and sell them 3 days later.
- Short-sell the 3 stocks with the lowest signals on the day of the news release and cover the short position 3 days later.
The initial cash is $1 million, which will be allocated equally to each trading day and each stock.
The performance will be assessed by the following metrics:
- Cumulative return. It is the total return generated by the trading strategy over a trading period.
- Win/loss ratio. It is calculated by dividing the number of winning trades by the number of losing trades over a trading period.
Participants need to submit the prompt and their LLM , which should be easily loaded and tested. Each team should also submit a 1-2 page report with the ACM sigconf template through Open Review. The title should start with “FinRL Contest 2024 Task II.”
For each task, the final ranking of participants will be determined by a weighted combination of model evaluation and report assessment, with weights of 60% and 40% respectively.
Model evaluation:
- Task 1: the geometric mean of the rankings of cumulative return, Sharpe ratio, and Max drawdown.
- Task 2: the geometric mean of the rankings of cumulative return and win/loss ratio.
Report assessment:
The assessment of the reports will be conducted by invited experts and professionals. The judges will independently rate the data and model analysis, results and discussion, robustness and generalizability, innovation and creativity, organization and readability, each accounting for 20% of the qualitative assessment.
Contestants can communicate any questions on Discord.
Contact email: finrlcontest@gmail.com