Skip to content

Reinforcement Learning for Trading Practical Examples and Lessons Learned by Dr. Tom Starke financial deepmind



This talk, titled, “Reinforcement Learning for Trading Practical Examples and Lessons Learned” was given by Dr. Tom Starke at QuantCon 2018.

Description:
Since AlphaGo beat the world Go champion, reinforcement learning has received considerable attention and seems like an attractive choice for completely autonomous trading systems. This talk shows practical aspects and examples of deep reinforcement learning applied to trading and discusses the pros and cons of this technology.

The slides for this talk can be viewed at:

About the Speaker:
Dr. Tom Starke has a Ph.D. in Physics and works as an algorithmic trader at a proprietary trading company in Sydney. He has a keen interest in mathematical modeling and machine learning in the financial markets. He has previously lectured computer simulation at Oxford University and lead strategic research projects for Rolls-Royce Plc.

Tom is very active in the quantitative trading community, running workshops for Quantopian, teaching people quantitative analysis techniques, and organizing algorithmic trading meetup groups such as Cybertraders Syd.

To learn more about Quantopian, visit

Disclaimer
Quantopian provides this presentation to help people write trading algorithms – it is not intended to provide investment advice.

More specifically, the material is provided for informational purposes only and does not constitute an offer to sell, a solicitation to buy, or a recommendation or endorsement for any security or strategy, nor does it constitute an offer to provide investment advisory or other services by Quantopian.

In addition, the content neither constitutes investment advice nor offers any opinion with respect to the suitability of any security or any specific investment. Quantopian makes no guarantees as to accuracy or completeness of the views expressed in the website. The views are subject to change, and may have become unreliable for various reasons, including changes in market conditions or economic circumstances. .

Images related to the topic financial deepmind

Reinforcement Learning for Trading Practical Examples and Lessons Learned by Dr. Tom Starke

Reinforcement Learning for Trading Practical Examples and Lessons Learned by Dr. Tom Starke

Search related to the topic Reinforcement Learning for Trading Practical Examples and Lessons Learned by Dr. Tom Starke

#Reinforcement #Learning #Trading #Practical #Examples #Lessons #Learned #Tom #Starke
Reinforcement Learning for Trading Practical Examples and Lessons Learned by Dr. Tom Starke
financial deepmind
See all the latest ways to make money online: See more here
See all the latest ways to make money online: See more here

19 thoughts on “Reinforcement Learning for Trading Practical Examples and Lessons Learned by Dr. Tom Starke financial deepmind”

  1. Thank you sir for good explanation!
    Please help me to solve this error – ImportError: cannot import name 'sgd' from 'keras.optimizers' am not able to fix this error and if anyone to fix this error please help me

  2. I am doing deep learning but now I'm thinking of integrating it with a reinforcement learning as an ensemble on the outside.
    (with a money management system on the side)
    Is there Anyone in California interested in my project?

  3. 30:14 you are updating state and applying the action. When we chose an action, first we need to apply than we need to update the state and get the reward. Let's say current price is 100.20. When agent decides to buy, it's has to buy from the price 100.20 (excluding spread/slippage and commission). In your example, it's buying with the next price. Am I wrong?

  4. I wonder how he smoothes the data – perhaps "now" timestamp was already including partial info of the next data point. If it was smoothly only backwards then the next timestamp at exit might be completely off than the real exit price.

  5. Just as with any other AI algorithm, you need to clean your data before you give it to your reinforcement learner. But you can make a neural net that cleans that data for you, with relative succes. Noise is also an issue in other domains, not just finance. Ofcourse you are creating a feedback loop. When you buy/sell with succes your competitors will adapt, and so the problem shifts to a more difficult state…adding overal noise (randomness) to the system.

Leave a Reply to Alexei Matusevski Cancel reply

Your email address will not be published. Required fields are marked *