Understanding Large Language Models: From Technical Function to Social Consequence

Our research project ‘SIxAI’ set out to explore how the continually expanding capabilities of artificial intelligence might be harnessed to augment the practice of systemic investing. But this objective conceals an assumption. By asking how we could use AI for SI, we skip over the question of whether AI should be used at all. After all, these are technologies mired in controversy. Are they really as powerful as many claim them to be? And even if they are, what risks do they pose, and what costs do they incur? Which agendas and worldviews do they advance? And which do they hinder or obscure? This article is our effort to engage with these questions, temporarily setting aside our belief that there is in fact exciting potential for AI to be applied to SI, in order to reflect more deeply on what AI really is, and how it should be approached and deployed in general, if at all. Our objective here is to ground all of our subsequent research in a robust and critical engagement with those thorny underlying philosophical questions of AI that often go un-asked. Here we present a series of 4 articles, introducing the technology itself, unpacking its capabilities and their limitations, discussing some of the uncomfortable truths underlying these capabilities, and finally proposing a set of principles, both for the responsible use of AI, and for its wise use too. As always, we welcome your comments.

 

PART 1: Understanding AI

Beyond the Hype

Is AI the saviour of the world, or its undoing? An accelerant of progress, or of chaos? You don’t have to look far in 2025 to find vocal proponents of either view. Indeed, “AI” is everywhere right now. Maybe it’s the spaces we spend time in, but it feels literally impossible to avoid the term wherever we go: every email newsletter, every social media post, every water-cooler-colleague-catch-up, every post-work-drinks-gathering seems incomplete without some mention of the term. Whatever your take, this is perhaps a sign that the ‘friend or foe’ question is now largely a rhetorical one, and that AI is here to stay regardless. Or, in the language of Roger’s Innovation Diffusion Theory, that “the late majority” have now come to terms with its usefulness, and that even “the laggards” are starting to resign themselves to its ubiquity, reluctantly filing into the party with the rest of us. But as “AI” assimilates into the cultural canon of twenty-first century life, what started as a precise scientific and philosophical term for a specific range of technological innovations is giving way to a hazy, catch-all term that gestures toward some inscrutable cliché (“clever tech stuff?”), with the majority of those who speak its name relating more to the sheer popularity of the field than to the inner workings of any particular tool. Much the same seems to have happened with “blockchain”, for example. There is a danger here, though: that we normalize the term so extensively that it seems to need no explanation, as though a harmless and vague fact of life, generic enough to contain any number of possible meanings without being individually accountable to any of them, permanently defanging our critique of how it really works, what it really represents, and whether or not it’s something we actually want in our lives. So, before we discuss what “AI” can and can’t do, what risks “it” does and does not pose, and whether or not we are witnessing its hand in the end of human civilization, let us first disambiguate. And apologies in advance to any readers with a highly sophisticated understanding of AI, for whom the summary below may be a cringe-worthy simplification: we welcome your corrections!

 

What is AI, though?

When people speak about “AI”, they’re typically referring to a variety of applications of “machine learning” (ML): a set of techniques that have been under development since the fifties, which enable computers to spot and contextualise patterns or rules in data without the need for humans to codify and input those patterns or rules beforehand. This is really at the core of all modern “AI” applications: autonomous pattern recognition. ML often uses a type of algorithm (an algorithm is essentially a set of instructions, like a recipe) known as “neural networks” (yes, modelled on the neurons of the human brain) to process the information needed to infer those patterns. But it’s advances in “deep learning” (which uses “deep neural networks” or DNNs, so called because they are structured into multiple stacked layers, huh) that are really causing all the fuss. DNNs – especially those using “transformer architecture” – are behind recent breakthroughs in, for instance, “natural language processing” (NLP), whereby computers can process linguistic data (giving a computer words that a human spoke or wrote and the computer determining what that human meant), and “computer vision” (CV), whereby computers can process visual data (giving a computer a picture of a train and the computer determining that it’s… a picture of a train) – both of which can be considered distinct subfields of AI.

Another subfield making waves right now is “generative AI” (GenAI). Essentially, GenAI refers to the creation and use of “large models” (LMs), which are trained to infer patterns from vast amounts of data, to generate “new” content, like text, images, sound, or video (whether or not it’s really “new” is a matter of spirited scholarly debate, but we’ll get to that later). With their algorithmic wizardry, “large models” (LM) condense immense amounts of complexity about, for example, how languages tend to work, or how images tend to look, into high-fidelity representations that can be used to generate predictions for what the next word in a given sentence might be, or what the missing patch in a given image might look like. After a few rounds of tweaking, an LM’s capacity for prediction can be used to generate endless reams of new content from a simple prompt. “Large language models” (LLMs), like those underlying household names such as ChatGPT, Claude, and Gemini, have typically been trained on tens (sometimes hundreds) of thousands of gigabytes of linguistic data, and their resulting proficiency in generating human-like language is already enough to fool humans into thinking they’re real people. But you probably already knew that bit.

LLMs: Under the Hood

In this research project, we’re using the term “AI” in reference to GenAI in particular, and especially to tools based on LLMs. So it’s worth going into a little more detail about how these models work. Let’s say you want to build an LLM: where would you start? Well, once you’ve developed your deep learning algorithm (no small feat), you need to expose it to data. A lot of it. This stage is known as basic training (although it can sometimes be preceded by an additional ‘pre-training’ step) and it provides the LLM with its modelling of grammar and semantics, its memorisation of basic world knowledge, and its discernment of human sentiment. The most common source of basic training data is known as “the Common Crawl”, which comprises an up-to-date repository of (almost) all of the freely available text on the open internet: every blog, every wikipedia entry, every Medium article, every angry Youtube comment thread. All of that text is first converted into strings of “tokens” (numbers which stand for clusters of letters that form the building blocks of all words, or “morphemes” if you’re a linguistics nerd) and then analysed for patterns. After this “unsupervised learning” period, an LLM will be able to call on millions of “parameters” (the factors and weightings it develops in order to model human language) to select the most plausible and contextually appropriate token to add to a given sequence, based on a dynamic probability distribution of all the possible options. So, if you were to feed the model a sentence beginning “the fox jumped onto the…”, the model would rank the probability of varying options for the next word. “Spaceship” would presumably be ranked very low; “skateboard”, just slightly higher but still fairly low; “cat”, quite highly; and “roof”, perhaps highest of all. And so it would select the term “roof”, and then move on. (It should be noted here, though, that how exactly an LLM generates its parameters and determines these probability distributions is nearly impossible to trace, a phenomenon often referred to as “the black box” or “explainability problem” of AI.) The sequence of tokens that precedes and informs a prediction is known as its “context window”. But these context windows aren’t necessarily just a few words long, as with our jumping fox example, but can also factor in the content of the preceding paragraphs, or even the sum total of all a user’s previous interactions with the model. The latest Gemini model released recently by Google, for example, has a maximum context window of one million tokens. That’s an amount of text equivalent to the entirety of Tolstoy’s War and Peace. One interesting thing to note here is that a model’s predictions are “autoregressive”, meaning that each new token that is selected and added to a given sequence then becomes part of the context window that’s used to predict the next token, and so on again and again. An LLM’s unsettlingly human-like linguistic proficiencies emerge, then, simply by virtue of its capacity to keep adding new tokens based on these probabilistic calculations. 

The next stage is known as fine-tuning, and consists of training the algorithm on datasets of sample prompts, so it develops the ability to read queues in the requests and questions likely to be posed by users and to adjust its responses accordingly, as well as conditioning it to elicit certain human-like qualities in those responses, like agreeableness or helpfulness. The result is an LLM that is either attuned to a specific use case, or that can adapt to a variety of types of task. Finally, your LLM is ready for “Reinforcement Learning from Human Feedback” (RLHF), whereby human workers label or score a range of outputs offered by the model in response to a given prompt based on certain parameters of desirability, for example that a clear answer is preferable to an ambiguous one, but that an ambiguous answer is preferable to one containing information that could threaten public safety, in order to further align the model’s behaviour with a user’s needs and expectations. Whip up a user interface that integrates the ability to search the web or call on curated databases of subject-specific expert knowledge alongside the core functions of the LLM, and there you have it: a market-leading AI product. Congratulations.

As you behold your creation, then, you may ask yourself: to which kinds of tasks should this “machine intelligence” be applied? And to which should we refrain from applying it? What inherent capabilities does it really possess, if any, and where do we meet the limits of those capabilities? And these would be very good questions. Check out the next blog in the series to discover (our take on) the answers!

Do you want to collaborate with us?

There is an urgent need to rethink the way we deploy financial capital for transformative impact in human and natural systems. The field of systemic investing has garnered significant momentum, and now is the time to scale deep and scale out. So we invite challenge owners, systems thinkers, innovation practitioners, investment professionals, ecosystem shapers, and creative voices to join us in figuring out how to redeploy financial capital in service of a prosperous and sustainable future for all.

How is systemic investing relevant to

Foundations

...because the pots of capital operating under a philanthropic logic are orders of magnitude smaller than those operating under an investment logic, so systemic investing is a way for foundations to leverage their capital in the systems they care about.

Corporations

...because their supply chains are becoming increasingly fragile and societal expectations of business are growing. This requires companies to deploy all the tools in their finance toolbox (incl. direct investments, advanced purchase agreements, and supply-chain financing) and partner more strategically with governments, foundations, and NGOs.

Impact Investors

...because single technologies, start-ups, or social enterprises—no matter how ingenious their solutions and how brilliant their teams—are unlikely to change systems by themselves. So what matters is that these single-point solutions are synergistically nested within a broader systems change effort.

Institutional Investors

...because mainstream ESG investing doesn’t benefit places and communities at the pace, scale, and quality required, so institutional investors must channel more capital into real-economy assets in a strategic and collaborative manner.

MDBs and DFIs

...because sustainable development in a VUCA world requires portfolio approaches to systems innovation, and those need to be funded with a different investment paradigm than those dominant in development finance institutions today. And because the public sector cannot finance sustainability transitions alone, so systemic investing is a way to crowd-in private-sector capital in a smart way.

Engage with us

Which option best describes your interest in systemic investing?

Your message has been sent!
Something went wrong – try again
About

Who We Are

The TransCap Initiative is a think-and-do-tank operating at the nexus of real-economy systems change, sustainability, and finance. We operate as a multi-stakeholder alliance coordinated by a backbone team and comprised of wealth owners, innovation leaders, system thinkers, research institutes, and financial intermediaries. Our community is open to anyone committed to our cause and values.

Why We Exist

We exist to improve the way sustainable finance is purposed, designed, and managed so that money can become a transformative force in building a low-carbon, climate-resilient, just, and inclusive society. We believe that the key to accomplishing this vision is to inspire and enable investors to leverage the insights and tools of systems thinking and complex systems science for addressing the most pressing societal challenges of the 21st century.

What We Do

Our mission is to build the field of systemic investing. This means developing, testing, and scaling an investment logic at the intersection of systems thinking and finance. We do that by convening a multi-stakeholder alliance to develop a knowledge and innovation base, test novel concepts and approaches, and build a community of practice.

Our core ideas borrow from the disciplines of systems thinking and complex systems science, challenge-led innovation, human-centred design, new economic frameworks, and financial innovation. Our experiments are contextualised in those place-based systems that matter most for human prosperity—such as cities, landscapes, and coastal zones—as well as in value chains and other real-economy systems. We hope that our work produces knowledge and insights, methods and tools, and a self-organising community of inspired and enabled change makers.

The places and value chains we intend to transform act as centres of gravity for our work. In each of these systems, we will work with challenge owners, communities, innovators, investors, and other stakeholders to design, structure, and finance strategic investment portfolios nested within a broader systems intervention approach.