ChatGPT is the fastest growing app of all time, gaining more than 100 million users just two months after its launch in November last year. It enables users to have human-like conversations that include reasonable-sounding and often correct answers to all sorts of questions. Like humans, it can ask for more information and explain reasoning.
The first academic research about the use of ChatGPT in finance is now being released. Two recent studies make GPT seem like a promising technology, to improve investment decisionmaking and to explain its decisions. Perhaps the long-held dream of replacing humans in finance is coming true.
In December last year, I wrote that “a tireless machine able to digest all information and immune to biases should be clearly superior to humans when it comes to investing. Except it’s not.”
Financial management was one of the earliest goals of artificial intelligence (AI) research because it seemed like an easy and highly rewarding task.
Yet so far, AI has succeeded only in niche applications in finance.
GPT stands for Generative Pre-trained Transformer, a five-year-old idea that might be a game-changer in AI applications.
Very broadly, there are three approaches to extracting useful information from data. With structured data, such as accounting numbers or price histories, statistics and formal models can be applied. With completely unstructured data — series of bits that could be photographs, physical measurements, text or anything else — there are algorithms that can extract patterns and predict inputs.
Language is somewhere in between. There is structure, meaning that only certain letter combinations are intelligible words and there are grammar rules for stringing words together.
Yet there are exceptions to rules and nuances beyond the literal text. A person needs a lot of domain knowledge and context to understand text.
There is an old story — it has been tracked back to 1956 at which time it was already old — about an AI worker who built a program to translate between English and Russian. She gave it the phrase “out of sight, out of mind” to translate to Russian, and then translated the Russian back to English and got “invisible idiot.”
There are no rules of language that tell us the phrase is an aphorism about forgetfulness rather than a description of an individual, but no native speakers would make the mistake.
GPT models are the hottest approach to working with language data, but quantitative trading and investment have used cruder language models for many years. A human researcher reads relevant information such as company statements, news stories, surveys and research reports carefully and slowly.
Computers can read vast quantities of information in many languages and come up with instant conclusions. This is essential for high-frequency trading when being a millisecond sooner to determine whether a news headline is good or bad news for a stock price is the name of the game.
Most of the language models used in quantitative finance today treat it as structured data. Algorithms look for certain words, or just measure the number of words in a headline or news release. Some algorithms look for certain patterns or structures, but none of the major ones try to understand the meaning of the text, and none of them can explain why they reach their conclusions or hold further conversation on the subject.
Now come two papers titled “Can ChatGPT Decipher Fedspeak?” and “Can ChatGPT Forecast Stock Price Movements?”
This is not a discussion about SkyNet taking over Wall Street, but whether ChatGPT beats older models — many of which treat language as structured — in making fast decisions about short texts.
The first paper asked ChatGPT to determine if an individual sentence from a US Federal Reserve statement was “dovish,” suggesting the central bank was more likely to cut than raise interest rates, or “hawkish,” suggesting the opposite.
A high-frequency trading algorithm might rate each sentence in the Fed release and use the output along with other data to trade federal funds futures or other instruments before the human analysts had finished reading the first word in the release.
In this study, ChatGPT clearly did better at matching conclusions of human analysts than dictionary-based models that looked only for certain words. When the researchers fine-tuned ChatGPT by giving it extra training on Fed statements with feedback on how humans rated the statements, it agreed with human researchers about as often as two human researchers agreed with each other, and its explanations for its decisions were plausible.
This is not immediately useful for trading. The paper did not disclose how fast the model ran, nor whether overall interpretations of entire Fed releases agreed well with human conclusions.
Whether they agreed with reality is not the point, as high-frequency traders are trying to beat the market to the new consensus, not to the theoretically correct place.
The paper said that GPT models might have turned a corner with regards to actually understanding language. If that is true — and one study does not prove anything — they could be unleashed on a much wider range of text to generate theses, such as whether inflation is likely to continue to be a problem over the next 12 months, rather than flash signals for high-frequency trading.
Instead of binary buy/sell signals, ChatGPT could also hold a conversation with a human analyst to improve investment decisions.
If this seems to be working, a future generation of GPT models could be trained on the entire history of texts and financial price movements.
The second paper is more directly relevant for trading. It used ChatGPT to rate news headlines as good or bad for stock prices. It tested the strategy of buying a stock with good news at the open after the headline was released and selling at the close; or selling at the open and buying back at the close if the headline was bad.
The results were inconclusive. The ChatGPT signal had a 0.01 correlation with the next day’s raw stock return.
However, to evaluate a signal one needs to compare to the residual return after adjusting for the market return, and perhaps for known factors. A 0.01 correlation could be valuable in combination with other signals, or it might not.
The tested strategy did have positive returns from October 2021 to December last year without transaction costs, but the authors do not provide data on whether it beat a market strategy, nor whether the positive return was significant statistically.
A reported 0.13 percent gross profit per trade suggests that it might not overcome transaction costs.
The authors also report a regression that includes future information, so it cannot be used to evaluate effectiveness for making decisions based on information known at the time.
The ChatGPT signal supplies no additional information to the three decimal places, although it does seem to have some small positive value, the authors said.
Yet inconclusive does not mean failure.
The study did say that ChatGPT was better than popular alternative models, and research on GPT and other large language models is continuing.
GPT is an AI tool that can work with humans, learn from them and teach them, rather than some incomprehensible black box. At the very least, it seems poised to replace older algorithms and to increase the use of AI in quantitative and qualitative investing.
It is a long way from taking over Wall Street, but there is no reason to think it cannot.
Aaron Brown is a former managing director and head of financial market research at AQR Capital Management. He might have a stake in the areas he writes about.
This column does not necessarily reflect the opinion of the editorial board or Bloomberg LP and its owners.
The image was oddly quiet. No speeches, no flags, no dramatic announcements — just a Chinese cargo ship cutting through arctic ice and arriving in Britain in October. The Istanbul Bridge completed a journey that once existed only in theory, shaving weeks off traditional shipping routes. On paper, it was a story about efficiency. In strategic terms, it was about timing. Much like politics, arriving early matters. Especially when the route, the rules and the traffic are still undefined. For years, global politics has trained us to watch the loud moments: warships in the Taiwan Strait, sanctions announced at news conferences, leaders trading
The saga of Sarah Dzafce, the disgraced former Miss Finland, is far more significant than a mere beauty pageant controversy. It serves as a potent and painful contemporary lesson in global cultural ethics and the absolute necessity of racial respect. Her public career was instantly pulverized not by a lapse in judgement, but by a deliberate act of racial hostility, the flames of which swiftly encircled the globe. The offensive action was simple, yet profoundly provocative: a 15-second video in which Dzafce performed the infamous “slanted eyes” gesture — a crude, historically loaded caricature of East Asian features used in Western
Is a new foreign partner for Taiwan emerging in the Middle East? Last week, Taiwanese media reported that Deputy Minister of Foreign Affairs Francois Wu (吳志中) secretly visited Israel, a country with whom Taiwan has long shared unofficial relations but which has approached those relations cautiously. In the wake of China’s implicit but clear support for Hamas and Iran in the wake of the October 2023 assault on Israel, Jerusalem’s calculus may be changing. Both small countries facing literal existential threats, Israel and Taiwan have much to gain from closer ties. In his recent op-ed for the Washington Post, President William
A stabbing attack inside and near two busy Taipei MRT stations on Friday evening shocked the nation and made headlines in many foreign and local news media, as such indiscriminate attacks are rare in Taiwan. Four people died, including the 27-year-old suspect, and 11 people sustained injuries. At Taipei Main Station, the suspect threw smoke grenades near two exits and fatally stabbed one person who tried to stop him. He later made his way to Eslite Spectrum Nanxi department store near Zhongshan MRT Station, where he threw more smoke grenades and fatally stabbed a person on a scooter by the roadside.