AI in asset management: A lightbulb moment
By Harry Moore; Martin Luk; Matthew Hertz, Man Group
Published: 23 September 2024
Today, it is difficult to imagine a world without power at the flick of a switch. Yet 20 years after the electric light bulb was invented by Thomas Edison in 1879, just 3% of US households had electricity. It took another two decades for mass adoption. This point is well made by Agrawal and his colleagues in ‘Power and Prediction’1 where they argue we are at a similar juncture in AI. We find ourselves in ‘The Between Times’, where there is plenty of enthusiasm while we await a truly game-changing application.
At Man AHL, we observe somewhat similar trends. Generative AI has certainly not yet replaced researchers or portfolio managers, or generated a whole new system for delivering market beating performance. What it has done, however, is boost productivity, allowing quantitative analysts (‘quants’) to spend more time focused on alpha generation. In this article, we showcase four examples of generative AI making an impact. We also discuss the challenges and opportunities of generative AI for the future of quant research.
Why the sudden uptick in AI hype?
The current focus is predominantly on generative AI. This leap allowing users to interact with models using human language and generate new outputs has been a significant driver of the recent excitement. Generative AI is a subset of machine learning, which is a subset of broader AI (Figure 1, below).
Figure 1. Subsets of Artificial Intelligence
For interested readers, the evolution of generative AI is covered in detail here.2 Machine learning techniques are already also well documented in asset management, with success enhancing asset predictors, improving risk management and driving down costs of execution.3 We have been using machine learning techniques at Man AHL4 for over a decade and are early adopters of generative AI.
Four ways we are using generative AI today:
We have a feeling generative AI will scale more quickly than the lightbulb of 1879. Our CIO recently detailed the adoption rate at Man Group5 (spoiler, it’s more than 70% of employees) while our Group CEO also discussed the efficiency gains here.6 Below we show four ways generative AI is making us more productive.
1. Coding with Copilot
One of the most effective use-cases of generative AI is assisting with coding. Tools like GitHub Copilot can accelerate the development of prototypes and initial research results by predicting code continuations, reducing development time. It also facilitates knowledge sharing, as developers can ask the AI to explain various parts of code written by others.
The challenge and the opportunity lie in training these tools to understand proprietary code. We have extensive code libraries for tasks like market data acquisition and running simulations. Off-the-shelf AI models lack knowledge of these specialised repositories.
We are developing chatbots with the capability to comprehend our internal code. For example, one chatbot can identify where to find metadata for a market code and retrieve prices, specifying the correct libraries and fields. It is a meaningful challenge which requires significant work to get useful outputs, but this capability enhances our efficiency and leverages our proprietary knowledge (Figure 2, below).
Figure 2. Copilot plotting a timeseries of S&P 500 E-Mini Future, using our internal libraries
2. Extracting information for trading catastrophe bonds
Man AHL was founded as a commodity trading advisor (CTA) trading futures contracts. Futures are highly standardised and liquid, making them easy to trade for a systematic investment manager. However, as we’ve grown and diversified our business, we increasingly trade more novel and exotic instruments.6
One example is catastrophe bonds, debt instruments designed to pay out when a pre-specified event occurs, typically a natural disaster. Each catastrophe bond has unique features which need to be clearly understood before investment and, unlike interest rate or credit default swaps, do not have standardised terms. This process involves reading the offering circular, which is done by a human analyst, and a second check of the extracted data, again by another human analyst. As these documents run to 200 pages, this can be a considerable amount of time (Figure 3, below).
Figure 3. Catastrophe bond offering circular
- Pages 200
- Words 50,000
- Expected human reading time 3 hours
Today, this data extraction is done by ChatGPT, putting the relevant information in a systematic template for a reviewer to check. This frees up one analyst to focus on new research.
3. Chatbots assisting with research for investor queries
Man Group’s Client Relations team assists with questions from clients on the firm’s full suite of investments, including Man AHL’s systematic investment strategies. Many questions rely on information from various investment materials, including factsheets, presentations, due diligence questionnaires, and investment commentaries. The team then crafts client-friendly responses. For example, a client might request information on fees, or the risk target of an investment strategy.
ChatGPT can automate several steps in this process (Figure 4, below). First, it can extract the required information from the relevant documents. Second, it can draft a response ready for human analyst review. This efficiency frees up time for the team to focus on higher-value tasks.
Figure 4. Screenshot of chatbot returning information
Improved efficiencies for data extraction are a general theme – we saw this in the catastrophe bond example – and other teams are reaping benefits too:
- Discretionary investment analysts extracting information from Regulatory News Service (RNS) announcements.
- Data science professionals extracting data information from vendor emails.
- Design professionals extracting underlying data from unformatted charts to convert to Man Group branding.
4. Producing simple macro hypotheses
ChatGPT is useful in quantitative macro research by leveraging its knowledge of fundamental macroeconomic relationships. One use-case is employing it as a hypothesis generator to suggest whether a particular economic timeseries has a fundamental relationship with a certain market. These hypotheses can then be tested using statistical back-testing methods.
While ChatGPT won’t replace our macro research team in its current state, its understanding can be as good as a graduate researcher. The main difference being that a human researcher needs breaks, while ChatGPT can query thousands of relationships systematically.
Figure 5. ChatGPT explaining a simple macroeconomic relationship
We’ve focused on the opportunities until now. Below we highlight some of the key lessons from our experience with generative AI.
Hallucinations must be managed
ChatGPT’s responses cannot be fully trusted. To help mitigate the impact of hallucinations, we use tools to highlight where information occurs in the original text, aiding human checking. It is a similar story for code, which is only a prototype and requires human verification.
Prompt engineering is crucial
If ChatGPT can’t do a task well, it’s often due to a misspecified prompt. Perfecting prompts requires significant resources, trial and error, and specific techniques.
Break tasks into smaller sub-tasks
ChatGPT can’t logically break down and execute complex problems in one go. Effective ‘AI engineering’ involves splitting projects into smaller tasks, each handled by specialist instances of ChatGPT with tailored prompts and tools. The challenge is integrating these agents to solve complex problems.
Education is key for wider adoption
Understanding ChatGPT’s capabilities and limitations is crucial. Sceptics should see its strengths, while enthusiasts need to learn its failures. Effective use requires learning how to interact with the model and understanding its training and functioning.
Be ready for the next best thing
Generative AI is already creating efficiencies in asset management, but users must be agile in taking on the next best model to reap the gains of this evolving technology. Progress has so far been swift, with GPT-2 released in 2019 described as ‘far from useable’7 versus GPT-4 which is already gathering multiple use-cases.
Conclusion
Electricity changed society but took 40 years to do it. AI can do the same,8 but faster. Our analysts have already seen improvements in data augmentation, feature engineering, model selection and portfolio construction. We believe that in as little as a decade, those asset managers who embrace generative AI can help gain a competitive edge via a faster pace of innovation and superior performance.
It was 40 years from Edison’s lightbulb until electricity changed the game for the masses. Perhaps AI hits that milestone in 10.
References
- Ajay Agrawal, A., Gans, J. and Goldfarb, A. ‘Power and Prediction: The Disruptive Economics of Artificial Intelligence’ (2022)
- Luk, M., ‘Generative AI: Overview, Economic Impact, and Applications in Asset Management’, 18 September, 2023. Available at: https://ssrn.com/abstract=4574814
- Ledford, A. ‘An Introduction to Machine Learning’, 2019. Available here: https://www.man. com/maninstitute/intro-machine-learning
- Korgaonkar, R., ‘Diary of a Quant: AI’, 2024. Available here: https://www.man.com/maninstitute/the-big-picture-ai
- Pensions and Investments, ‘Man Group CEO sees generative AI boosting efficiency, but not investment decisions’, 30 April 2024. Available here: https://www.pionline.com/hedgefunds/generative-ai-has-boosted-efficiency-not-investment-decision-making-mangroup-ceo-robyn
- Korgaonkar, R., ‘Diary of a Quant: Journeying into Exotic Markets’, 2024. Available here: https://www.man.com/maninstitute/the-big-picture-journeying-exotic-markets
- Radford, A., et al., ‘Language models are unsupervised multitask learners’, 2019. OpenAI blog, 1(8), p.9
- Eloundou, T., et al., 2023. ‘Gpts are gpts: An early look at the labor market impact potential of large language models’. arXiv preprint arXiv:2303.10130