Artificial intelligence has moved quickly from experimentation to execution. As models become more capable, cheaper to deploy, and embedded directly into workflows, AI is no longer a futuristic concept but a practical tool reshaping how companies build, operate, and compete. The focus is shifting away from novelty and towards impact.
In this article, our partners at Top Venture take a closer look at why AI is now “locked and loaded” for its next phase of growth. Moving beyond the initial hype, they explore where real adoption is taking place, how business models are evolving, and what this could mean as the theme matures.
AI Locked and Loaded
Take a minute to think about a recent conversation you had – perhaps at a dinner party – when you were asked a question on a topic you cared about or knew a lot about. Because you had already given that subject considerable thought, you may have been locked and loaded, ready to deliver a beautiful, long, and coherent answer. As you spoke, each sentence flowed naturally into the next. Each word, each conjugation, seemed to appear without conscious effort. You were on a roll!
In such moments, you are not deeply planning each sentence in advance, nor carefully selecting every word before it comes out. Most often, you are not constructing your argument through a deliberate logical progression before you begin speaking. Instead, each sentence emerges fluidly. Without thinking explicitly about each word, your brain’s neural network draws on years of accumulated knowledge and on grammar rules drilled into your head since kindergarten, allowing you to produce coherent speech almost automatically.
This is a useful analogy for how the current generation of generative AI works.
AI on a Roll
Large language models (LLMs) such as ChatGPT, Gemini and Claude also generate their responses one word at a time in a similar progression. To enable this, their creators – leading LLM companies like Anthropic, Google, and OpenAI – spend enormous sums training each version of their models. During training, grammar rules, sentence structures, and conceptual associations are implicitly encoded into the LLM’s neural network as weights: parameters that determine how likely each word is to follow a given sequence of words.
Once this long and expensive training phase is complete, LLM companies can release the model to the public because, in effect, it is locked and loaded. When you ask it a question, it is ready to string together the next word and the next sentence based on those probabilities, producing a coherent response without explicitly “thinking” about each step. In technical jargon, this is called autoregressive generation. In practice, we see that the LLM, too, is on a roll!
AI off the Rails
This analogy also sheds light on how LLMs can hallucinate – those moments when a model suddenly goes off the rails and starts making things up. Think back to a time when you were on a roll, talking fluidly about a subject, and then made a verbal gaffe: saying something you really shouldn’t have said simply because your brain associated a particular idea or phrase with the conversation at hand. If you had paused to think more carefully, you likely would have caught yourself – but in the moment, the words just came out, embarrassing you, simply because they were associated with one another in your brain.
Similarly, LLMs do not inherently evaluate whether the content they are generating is logically correct or factually grounded. When they hallucinate, they are often just following probabilistic associations learned during training, pushed along by the momentum of their own output. In other words, because the LLM is using a probability table to come up with every word, each time you ask it a question, its answer will be slightly different. So, there is always a chance that sometimes, the next word or sentence it chooses is one of the less probable ones, perhaps an awkward one, like a gaffe or a falsehood. What’s worse, once they state that gaffe or falsehood, the LLM may continue generating sentences associated with that awkwardly improbable sentence, sliding down a slippery slope of extended hallucinations.
If you have used LLMs regularly over the past year or two, however, you may have noticed that these hallucinations are becoming less frequent – especially in the most recent generations of so-called reasoning models, which have been trained to check themselves. One way humans reduce gaffes is by becoming more deliberate: performing a quick reality check on what we are about to say and stopping ourselves before we articulate something incorrect or embarrassing.
Recent LLM systems are trained to approximate this behavior by incorporating mechanisms that encourage similarly deliberate reasoning. They examine their internal chain-of-thought for errors and embarrassing derivations before committing to a final answer. As a result, these models increasingly behave less like unfiltered dinner-party guests and more like careful participants in measured conversations or business negotiations.
LLM Education
Another way humans improve the quality of their arguments is through preparation. Before an important meeting or a high-school debate, we study the subject, rehearse responses, and do dry runs to learn from our mistakes. With sufficient preparation, we can often respond coherently even to questions we did not explicitly anticipate. A good education and broad exposure to knowledge make this possible.
The training of large language models has proven to follow a similar pattern – except that LLMs are not preparing for a single debate or meeting. They are being trained to respond to billions of queries across virtually every subject imaginable, often from users actively trying to probe their weaknesses. One of the most fascinating observations of LLM researchers is that the more knowledge an LLM possesses, the more capable it appears to generalize that knowledge, and to apply it to areas it was not even specifically trained on.
Just as teaching children a broad curriculum – from math and science to literature and the arts – prepares them to reason more effectively, LLMs seem to benefit from broad training even when tackling narrow tasks.
Data Quality vs Quantity
This insight – that scaling data, compute, and model size could dramatically improve performance – was not obvious at first and surprised even leading researchers. Scientists such as Dario Amodei, now the founder and CEO of Anthropic, began exploring these scaling laws around a decade ago while on early language models at OpenAI. As models were trained on increasingly large datasets, often encompassing much of the publicly accessible internet, their capabilities continued to improve. When high-quality human-generated data became scarce, researchers supplemented it with synthetic data and increasingly sophisticated training techniques.
Throughout this process, skeptics – including many AI researchers – repeatedly questioned whether continued scaling would yield meaningful gains. Thus far, scaling has consistently delivered improvements that users experience firsthand when interacting with modern LLMs, even as debates continue about how long this trend can persist.
Leaders in the field, including Google Gemini’s Demis Hassabis, openly acknowledge that gains from additional compute are showing diminishing returns. Amodei himself does not rule out the possibility that progress from pure scaling could eventually plateau. At the same time, there is broad consensus among those at the forefront of the technology – including Hassabis and Amodei – that no such hard limit is currently visible and that meaningful improvements are likely to continue for the foreseeable future.
One reason for this optimism is that scaling is no longer just about ingesting more raw data. Leading LLM companies now invest heavily in improving data quality, generating high-quality synthetic data, and building feedback loops that help models distinguish better answers from worse ones. To extend the school analogy, these companies are not just giving their LLM students more books; they are providing better textbooks, writing proprietary materials, and administering increasingly sophisticated exams – many, many exams. Unsurprisingly, the quality of the LLM’s school – and its teachers, has had a significant impact on model performance.
Think Broader, not Bigger
Despite all the talk about ever-larger data centres, it is generally understood that improvements in modern LLMs cannot be attributed to any single breakthrough. In one illustrative experiment, Andrej Karpathy – an early Tesla and OpenAI researcher and one of the sages of the modern LLM wave – re-ran one of the first image-recognition models using the original data and later applied modern algorithms to it. He found that the error rate was roughly halved purely due to improved techniques, before adding any additional data. Only afterward did he further improve performance by giving the model ten times more data.
This experiment reinforces the thesis that progress in model performance has come, and will likely continue to come, from many relatively small innovations across architecture, data curation, optimization, training methods, evaluation, and hardware: incremental advances that compound over time.
How, then, do companies create this steady stream of innovations? How can they keep improving their schools and teachers? Largely by attracting exceptional talent. This explains why competition for elite AI researchers has intensified dramatically, with reports of extraordinarily large compensation packages being offered to recruit top scientists. Yet at the highest levels, financial incentives alone have proven insufficient. Many leading researchers choose environments where they can collaborate with peers they respect and learn from others at the cutting edge. It turns out that what attracts talent most is talent itself.
This dynamic helps explain why organizations such as Google’s Gemini and Anthropic have become powerful talent magnets. It also sheds light on why technology giants like Amazon, Microsoft, and Apple have often chosen partnerships over direct competition by partnering with companies like OpenAI and Anthropic.
In brief, the leadership positions of OpenAI and Anthropic are more defensible than they initially appear. These early AI leaders have a strong first-mover advantage, due to their talent density, the know-how and research they have built on this foundation, the feedback loops their millions of users provide for continuous improvement, the distribution channels they are building, and finally, the efficiency they gain from scale.
This is the first of four posts about the current state of AI. In the second article, we will delve more deeply into the competitive and strategic dynamics of the industry and the potential threats to the current leaders in this field.
To learn more about our approach or explore how we partner with multi‑generational families, please contact: info@bedrockgroup.com.
Important Legal Information
The views presented in this article are Salman’s own views and might not necessarily reflect the investment approach of Bedrock Group.
The information and opinions contained in this article are for background information and discussion purposes only and do not purport to be full or complete. No information in this document should be construed as providing financial, investment or other professional advice. The views and opinions expressed herein are based on current market conditions and are subject to change without notice. No representation is made that any forecast or projection will be realised.
Information included in this document is intended for those investors who meet the definition of Professional Client under the Swiss FinSA regulation as well as Professional Client or Eligible Counterparty under the UK Financial Conduct Authority.
Investment Risks
The value of all investments and the income derived therefrom can fluctuate due to market movements and you may not get back the amount originally invested. invested.
Limitation of Liability and Indemnity
Bedrock expressly disclaims liability for errors or omissions in the information and data contained in this document. No representation or warranty of any kind, implied, expressed or statutory, is given in conjunction with the information and data. Bedrock accepts no liability for any loss or damage arising out of the use or misuse of or reliance on the information provided including, without limitation, any loss of profits or any other damage, direct or consequential. You agree to indemnify and hold harm less Bedrock and its affiliates, and the directors and employees of Bedrock and its affiliates from and against any and all liabilities, claims, damages, losses or expenses, including legal fees and expenses arising out of your access to or use of the information in this presentation, save to the extent that such losses may not be excluded pursuant to applicable law or regulation. Any opinions contained in this presentation may be changed after issue at any time without notice.
After spending a week inside Silicon Valley’s AI ecosystem, David Ephrati argues that today’s AI moment is not a repeat of the internet wave, but something materially larger. This article distils seven observations he believes carry lasting relevance for investors.
Gain firsthand insights from Silicon Valley’s top visionary founders, investors, and industry leaders on the AI, robotics, and venture capital trends shaping the valley.
In this article, we explore some historical examples of the types of companies that have benefited from new disruptive technologies - notably Coca-Cola's transformation as a result of refrigeration - and draw some conclusions about the potential impact of AI, highlighting some of the companies set to be potential winners.