ChatGPT is the fastest growing app of all time, gaining more than 100 million users just two months after its launch in November last year. It enables users to have human-like conversations that include reasonable-sounding and often correct answers to all sorts of questions. Like humans, it can ask for more information and explain reasoning.
The first academic research about the use of ChatGPT in finance is now being released. Two recent studies make GPT seem like a promising technology, to improve investment decisionmaking and to explain its decisions. Perhaps the long-held dream of replacing humans in finance is coming true.
In December last year, I wrote that “a tireless machine able to digest all information and immune to biases should be clearly superior to humans when it comes to investing. Except it’s not.”
Financial management was one of the earliest goals of artificial intelligence (AI) research because it seemed like an easy and highly rewarding task.
Yet so far, AI has succeeded only in niche applications in finance.
GPT stands for Generative Pre-trained Transformer, a five-year-old idea that might be a game-changer in AI applications.
Very broadly, there are three approaches to extracting useful information from data. With structured data, such as accounting numbers or price histories, statistics and formal models can be applied. With completely unstructured data — series of bits that could be photographs, physical measurements, text or anything else — there are algorithms that can extract patterns and predict inputs.
Language is somewhere in between. There is structure, meaning that only certain letter combinations are intelligible words and there are grammar rules for stringing words together.
Yet there are exceptions to rules and nuances beyond the literal text. A person needs a lot of domain knowledge and context to understand text.
There is an old story — it has been tracked back to 1956 at which time it was already old — about an AI worker who built a program to translate between English and Russian. She gave it the phrase “out of sight, out of mind” to translate to Russian, and then translated the Russian back to English and got “invisible idiot.”
There are no rules of language that tell us the phrase is an aphorism about forgetfulness rather than a description of an individual, but no native speakers would make the mistake.
GPT models are the hottest approach to working with language data, but quantitative trading and investment have used cruder language models for many years. A human researcher reads relevant information such as company statements, news stories, surveys and research reports carefully and slowly.
Computers can read vast quantities of information in many languages and come up with instant conclusions. This is essential for high-frequency trading when being a millisecond sooner to determine whether a news headline is good or bad news for a stock price is the name of the game.
Most of the language models used in quantitative finance today treat it as structured data. Algorithms look for certain words, or just measure the number of words in a headline or news release. Some algorithms look for certain patterns or structures, but none of the major ones try to understand the meaning of the text, and none of them can explain why they reach their conclusions or hold further conversation on the subject.
Now come two papers titled “Can ChatGPT Decipher Fedspeak?” and “Can ChatGPT Forecast Stock Price Movements?”
This is not a discussion about SkyNet taking over Wall Street, but whether ChatGPT beats older models — many of which treat language as structured — in making fast decisions about short texts.
The first paper asked ChatGPT to determine if an individual sentence from a US Federal Reserve statement was “dovish,” suggesting the central bank was more likely to cut than raise interest rates, or “hawkish,” suggesting the opposite.
A high-frequency trading algorithm might rate each sentence in the Fed release and use the output along with other data to trade federal funds futures or other instruments before the human analysts had finished reading the first word in the release.
In this study, ChatGPT clearly did better at matching conclusions of human analysts than dictionary-based models that looked only for certain words. When the researchers fine-tuned ChatGPT by giving it extra training on Fed statements with feedback on how humans rated the statements, it agreed with human researchers about as often as two human researchers agreed with each other, and its explanations for its decisions were plausible.
This is not immediately useful for trading. The paper did not disclose how fast the model ran, nor whether overall interpretations of entire Fed releases agreed well with human conclusions.
Whether they agreed with reality is not the point, as high-frequency traders are trying to beat the market to the new consensus, not to the theoretically correct place.
The paper said that GPT models might have turned a corner with regards to actually understanding language. If that is true — and one study does not prove anything — they could be unleashed on a much wider range of text to generate theses, such as whether inflation is likely to continue to be a problem over the next 12 months, rather than flash signals for high-frequency trading.
Instead of binary buy/sell signals, ChatGPT could also hold a conversation with a human analyst to improve investment decisions.
If this seems to be working, a future generation of GPT models could be trained on the entire history of texts and financial price movements.
The second paper is more directly relevant for trading. It used ChatGPT to rate news headlines as good or bad for stock prices. It tested the strategy of buying a stock with good news at the open after the headline was released and selling at the close; or selling at the open and buying back at the close if the headline was bad.
The results were inconclusive. The ChatGPT signal had a 0.01 correlation with the next day’s raw stock return.
However, to evaluate a signal one needs to compare to the residual return after adjusting for the market return, and perhaps for known factors. A 0.01 correlation could be valuable in combination with other signals, or it might not.
The tested strategy did have positive returns from October 2021 to December last year without transaction costs, but the authors do not provide data on whether it beat a market strategy, nor whether the positive return was significant statistically.
A reported 0.13 percent gross profit per trade suggests that it might not overcome transaction costs.
The authors also report a regression that includes future information, so it cannot be used to evaluate effectiveness for making decisions based on information known at the time.
The ChatGPT signal supplies no additional information to the three decimal places, although it does seem to have some small positive value, the authors said.
Yet inconclusive does not mean failure.
The study did say that ChatGPT was better than popular alternative models, and research on GPT and other large language models is continuing.
GPT is an AI tool that can work with humans, learn from them and teach them, rather than some incomprehensible black box. At the very least, it seems poised to replace older algorithms and to increase the use of AI in quantitative and qualitative investing.
It is a long way from taking over Wall Street, but there is no reason to think it cannot.
Aaron Brown is a former managing director and head of financial market research at AQR Capital Management. He might have a stake in the areas he writes about.
This column does not necessarily reflect the opinion of the editorial board or Bloomberg LP and its owners.
Congratulations to China’s working class — they have officially entered the “Livestock Feed 2.0” era. While others are still researching how to achieve healthy and balanced diets, China has already evolved to the point where it does not matter whether you are actually eating food, as long as you can swallow it. There is no need for cooking, chewing or making decisions — just tear open a package, add some hot water and in a short three minutes you have something that can keep you alive for at least another six hours. This is not science fiction — it is reality.
A foreign colleague of mine asked me recently, “What is a safe distance from potential People’s Liberation Army (PLA) Rocket Force’s (PLARF) Taiwan targets?” This article will answer this question and help people living in Taiwan have a deeper understanding of the threat. Why is it important to understand PLA/PLARF targeting strategy? According to RAND analysis, the PLA’s “systems destruction warfare” focuses on crippling an adversary’s operational system by targeting its networks, especially leadership, command and control (C2) nodes, sensors, and information hubs. Admiral Samuel Paparo, commander of US Indo-Pacific Command, noted in his 15 May 2025 Sedona Forum keynote speech that, as
In a world increasingly defined by unpredictability, two actors stand out as islands of stability: Europe and Taiwan. One, a sprawling union of democracies, but under immense pressure, grappling with a geopolitical reality it was not originally designed for. The other, a vibrant, resilient democracy thriving as a technological global leader, but living under a growing existential threat. In response to rising uncertainties, they are both seeking resilience and learning to better position themselves. It is now time they recognize each other not just as partners of convenience, but as strategic and indispensable lifelines. The US, long seen as the anchor
Kinmen County’s political geography is provocative in and of itself. A pair of islets running up abreast the Chinese mainland, just 20 minutes by ferry from the Chinese city of Xiamen, Kinmen remains under the Taiwanese government’s control, after China’s failed invasion attempt in 1949. The provocative nature of Kinmen’s existence, along with the Matsu Islands off the coast of China’s Fuzhou Province, has led to no shortage of outrageous takes and analyses in foreign media either fearmongering of a Chinese invasion or using these accidents of history to somehow understand Taiwan. Every few months a foreign reporter goes to