An external "AI brain." Big swings in biopharma. Infinite games. A nuclear resurgence. "Faceless" creators. Google search challengers. Battlefield AI. We asked 50 a16z partners to preview one big idea that will spur innovation in 2025.
In 2025, I believe we’ll see a surge in demand for nuclear energy. A perfect storm of regulatory reform, public enthusiasm, capital infusions, and insatiable energy needs — particularly from AI data centers — will accelerate orders for new reactors for the first time in decades.
As AI advances, America’s energy demand is skyrocketing. For the first time in decades, electricity consumption is on the rise, rattling our aging grid and reigniting the search for new, reliable power sources. Hyperscale data centers, hungry for clean and consistent energy, are already reviving decommissioned nuclear plants, including Pennsylvania’s once-infamous Three Mile Island, slated to come back online in 2028.
Bipartisan momentum and grassroots support for clean energy have spurred renewed interest in nuclear power. But this is about more than energy: it’s about securing America’s leadership in the global AI race, building a more resilient grid, and future-proofing national prosperity.
In the 2000s and 2010s, if you weren’t coding, it seemed like you’d get left behind. The number of Computer Science majors exploded, while degree programs like Mechanical Engineering and Electrical Engineering shrunk, on a relative basis.
Now we’re beginning to see a crucial shift into jobs that make the latest advancements in AI usable in complex hardware contexts. Amid the push to reshore manufacturing; the mass retirement of skilled workers across unsexy industries like water treatment, commercial HVAC, and oil and gas; and the rise of autonomy across defense, enterprise, and consumer applications, we’re seeing a renaissance in technical disciplines that cross the hardware-software chasm. These are the new Jobs of the Future.
In particular, I expect that 2025 will see accelerated demand for: electrical engineers, controls engineers, mechanical and mechatronics engineers, manufacturing engineers, RF engineers, industrial engineers, test engineers, quality engineers, and high-skilled technicians/robotic teleoperators of all types. The growth in some of these sectors may even outpace that of “traditional” software engineering over the next decade. The robots are coming — someone will have to build, train, and service them.
With the “catch” of the Starship booster bringing us one step closer to full rapid reusability of a 150+ metric tons capable lift vehicle, we could be approaching a new era in space. Starship’s success advances the ability to send humans to the Moon and Mars by lifting data centers, space stations, and other large capacities into space. It means we could have the capacity to develop biomedical labs in microgravity environments. And we’ll be able to transport humans and cargo around planet Earth in 40 minutes or less, opening up unprecedented industries and access. 2025 could be the beginning of fiction becoming reality.
I expect military decisions will increasingly be made at the edge, where autonomous drones, sensor networks, and battlefield AI systems operate with minimal human intervention. Imagine real-time tactical adjustments happening in remote conflict zones, where soldiers rely on AI-driven insights and automated systems to make split-second decisions. Small, mobile command centers may be positioned in hostile terrain, processing vast amounts of data locally to avoid latency and the risks of centralized control. Critically, this shift demands reliable, scalable compute and electrical power in remote and hostile environments to power advanced weapon systems. It’s not just a logistical challenge; it’s a supply chain and technological revolution.
To maintain operational superiority, we must invest in technologies that enable rapid force projection even in contested environments. Energy, space, and AI advancements can all play a critical role in hardening our capabilities against our international rivals. The future of defense is likely to be decentralized, and the race to secure the edge may define the next generation of military innovation.
I believe XR devices will become a valuable tool in enabling developers to build physical world applications. A new generation of extended reality platforms emerged this past year, including Apple’s Vision Pro and Meta’s Orion augmented reality glasses. While these products are still early in terms of consumer adoption and developer activity, they have great potential for unlocking more capable physical world uses, particularly when applied to advancing fields like world models and robotics foundation models.
Already, these devices play an important role in robotics, autonomy, and simulation. The implications of widely available XR devices — coupled with increased developer activity on spatial computing platforms — are particularly promising in verticals that involve a lot of physical world data, interactions, and infrastructure.
The number of Earth observation satellites has doubled in the last five years, from 500 to over 1,000. There is more data than ever downlinked to the Earth, and it is easier than ever to access imagery (although it is still far too difficult). Sophisticated government and academia customers have devoted considerable resources to building tools that rely on Earth observation data to meet their needs.
Yet commercial revenue in the sector still has not materialized in a meaningful way.
The real commercial opportunity lies in developing verticalized products that use Earth observation as just one tool among many to address industry-specific needs. These products should actually solve practical customer problems — considering labor shortages, budget constraints, and other business challenges — rather than simply providing another analysis or dashboard tool. In the future, latent Earth observation data could enhance the work of every supply chain manager, urban planner, and first responder.
Many companies are already investing in large-scale robot data collection via teleoperation, synthetic sim2real, modular manipulator attachments, and other embodiments. These methodologies can help collect the foundation-scale data necessary for generalized robots in the real world. Once we’ve collected billions or trillions of tokens of robot data, however, what comes next?
Scale AI rose to prominence during the self-driving boom by labeling data for perception: drawing boxes around cars and segmenting the road from the sky. Today, AI advances have yielded new forms of data collection no longer directly connected to the exact policies that model-builders are trying to train. The gen AI boom has ushered in a new focus on benchmarks, preference data, and safety and red-teaming — no longer simply labeling and collecting the exact policy data, but moving up the value chain to more complex, expensive, and difficult tasks.
What does this look like for robotics? Will we see warehouses of robots posed in hazardous environments for safety evaluations? Will organizations create complex, hidden benchmarks to determine vertical-specific winners? While we definitely need organizations collecting data to train the first truly generalizable robot policies, actually deploying those policies will rely on these second-order systems.
In 2025, I expect continued advancements in free-space optical communications, driven by nearly inelastic demand for communications bandwidth. Free-space optical links offer significant advantages over traditional radio frequency methods, particularly in the form of higher bitrates and higher transmission directivity. However, current optical communication technologies remain rudimentary in their efforts to mitigate downtime and interference, often relying heavily on transport layer protocols such as TCP for error correction.
Exciting developments include the introduction of more advanced modulation schemes, similar to innovations like QPSK and OFDM in wireless communications, which enhance data transfer efficiency. Additionally, we expect to see improvements in resiliency, with better beam steering and control, as well as more effective error correction to mitigate environmental factors. The integration of more precise positioning, navigation, and timing (PNT) systems will also enhance optical communication by enabling more accurate beam alignment, particularly in mobile applications. As progress is made, these advances will impact telecommunications, satellite communications, and our defense capabilities.
In 2025, “big is back” in the biopharma world, as even early-stage biotech startups begin to go after big, common diseases again.
What’s driving the shift? GLP-1 drugs for diabetes and obesity are expected to create a $100B+ market by 2030, and have infused new energy into the cardiometabolic disease space. Perhaps more quietly, we are also experiencing a gradual revolution in how we understand and treat many common auto-immune diseases, like lupus and arthritis. Recently, a German physician in Munich named Dr. Georg Schett hypothesized that engineered CAR-T cell therapy used for certain B cell cancers might also help patients with B-cell-driven auto-immune diseases — and this year, he published astonishing results. In a cohort of 15 patients for whom no other therapies were offering benefit, every single patient experienced dramatic improvement. As Dr. Schett described it: “The CAR-T therapy is like a reset button on a computer; it basically restarts the system and the immune system works perfectly fine.”
Inspired by powerful results like this, as well as the clinical and commercial success of new obesity medications, we expect a new wave of biotech and startup innovation focused on treating (and potentially even curing) our biggest diseases.
The last several years have seen the democratization of health through innovative tech options, from AI that analyzes blood biomarkers, to wearables that track biometrics, to accessible full-body screening. These tools empower patients to take control of their own health, giving them unprecedented access to personal data and insights that were once only available in clinical settings. AI plays a crucial role in this transformation, providing deeper analysis, personalized recommendations, and early detection, while identifying patterns that might otherwise go unnoticed.
In a difficult-to-navigate, reactive health system that’s more focused on “sick care” than healthcare, this democratization of health is revolutionary. By enabling people to monitor their health proactively, these tech options refocus the healthcare model on prevention, early detection, and keeping healthy patients healthy for longer. The result is a more patient-centered system, with the power of predictive insights allowing individuals to make informed, timely decisions about their health.
Healthcare is dealing with the mother of all clinical staffing crises — we’re short hundreds of thousands of doctors and nurses relative to the level of (rapidly growing) demand for clinical services that is projected in the next 5 years. On the administrative side, we’re facing the mother of all staffing paradoxes, which is that we use too many humans to perform basic, rote work that adds unnecessary cost to our already bloated system.
One of the most profound challenges we face in healthcare is how to get the most leverage out of the clinical and administrative workers we have, while using technology to systematically automate tasks that are below the pay grade of the humans who perform them today.
One of the most profound technologies that has presented itself to the industry to take on this challenge is AI. We believe 2025 will see specialist AI models serving as “super staffing” platforms in the high stakes environment of healthcare. Companies doing so will be able to start tapping into labor budgets vs. IT budgets to unlock an order of magnitude more scale of opportunity than the last era of traditional healthcare IT.
Making medicines is hard; biology is incalculably complex. The goal is to identify potential disease targets for drugs and validate them through rigorous experimentation. The ultimate confirmation, FDA approval, can take a decade and a billion dollars.
The prize is worth the pursuit. New medicines generate massive value for patients, companies, and society. But timing is everything. Developing a drug against the wrong target is an expensive failure; missing out on the next GLP-1 is a costly mistake.
Biological targets are like avocados: too early… too early… too early… now — too late! The emergence of appetizing targets can be unpredictable. Once a target is validated, a feeding frenzy ensues. Companies race to make drugs against it — now with heightened intensity, as a ravenous China enters the fray and has large pharmas’ and investors’ attention.
So what’s a startup to do? Looking to 2025, startups increasingly need an earned secret: a unique insight (head start) around an emerging target; a differentiated approach to hitting a hot one. Technology and AI are powerful tools for earning, and keeping, a secret.
Startups will need to be first, furtive, and fast to scoop up ripening avocados — or they’re toast.
Imagine being in a band with an AI drummer: every time you switch up the tempo or spontaneously introduce a new riff, the beat changes in unison, matching you note-for-note. It’s seamless and intuitive, meeting your ability, complementing your vibe, and keeping pace better than your usual halting jam sessions. The emergence of real-time AI brings that vision closer to reality.
We saw the first glimmers of near real-time AI with the release of Latent Consistency Models (LCMs) in 2023. As the inference time continues to drop, AI creative tools have become both more productive — yielding more generations per second — and more useful, giving way to new use cases like live video-to-video. In the coming year, exciting new use cases will be unlocked: everything from generated video companions to AI bandmates.
As latency disappears, new possibilities emerge. This technology even has potential in educational settings, giving teachers the ability to shift gears in a lesson or restate a point based on real-time micro-responses, such as students appearing distracted or confused. An immediate feedback loop allows us to prototype, iterate, and refine ideas at breakneck speed. This shift will redefine every creative workflow, where experiencing ideas as they happen unlocks the prospect of true co-creation with machines.
Already, anyone can create realistic video clips from a single image or simple text prompt. Several products have emerged in this space over the past two years offering comparable functionality, but diverging levels of consistency and quality. In 2025, I expect AI-generated video to become further segmented by use case, giving creators more control and better results.
In the year ahead, I anticipate AI video tools will achieve greater depth in storylines (stretching beyond context-free, five-second clips), higher quality and character consistency, and — most importantly — increased specialization. Video generation models will be trained to specific uses: for product marketing; for cinematic, long-form film; for hyperrealistic, 3D avatars; for seamless background images and B-roll; for anime conversion. They’ll become optimized for particular channels, as well, whether TikTok, YouTube, ad campaigns, or the big screen.
There are giant companies to be built from every seemingly niche video tool. In the coming years, AI video will evolve from impressive prototype to art form.
We all produce a ton of digital exhaust through our text messages, emails, tweets, browsing history, TikTok/Reddit comments, and more. Thanks to LLMs, we can now put all of this unstructured data to use in a “digital brain” that understands how you think and feel.
This isn’t just science fiction. I spent six months “exporting” my own brain to ChatGPT, and I was blown away by how good it was at helping me navigate both personal and professional situations.
I expect there will be many use cases and applications for this. There will be apps that help you better understand yourself, products that guide you in communicating with others, and tools that make you more effective at work. Just as LLMs can take massive amounts of information, extract the insights, and summarize the takeaways, AI-powered apps can keep a record of our thoughts, like a digital journal. With emerging consumer products tailored to this purpose, in the year ahead I predict more people will begin using AI as an infinite memory bank to guide their decision-making, interactions, and personal growth.
AI is great at producing things, but it isn’t great at producing things that actually sound like you. As anyone who’s experimented with AI writing knows, a bad draft can be worse than none at all. Style and tone often make the difference between a usable draft and something that requires heavy editing.
In images, we’ve already seen the rise of LoRAs and Midjourney style references (SREFs), which allow users to control the style and look of the output. I’m excited to see similar controls come to knowledge work. Beyond basic autocomplete, how can AI write an email that really sounds like you? How can AI retrieve information for, create, and format a slide in a deck in the way your company requires?
There are many ways to solve this problem, and the solution may look different based on the role and work product. In some cases, this may even involve the AI working as a copilot, “tapping in” the human when it needs an assist or information. Not everything will be one-shot, from prompt to full output. This feels crucial in moving toward a world where everyone does a significant part of their work with AI every day.
We’re approaching a qualitative data revolution. Historically, analytical software has been confined to numbers and structured data, which only represents a sliver of the greater story. Spreadsheets are effective for quantitative tasks, but the big picture is derived from words, narratives, and unstructured insights.
With the emergence of LLMs, web-based agents, and multimodal models, we can now collect, comprehend, and integrate unstructured data with quantitative information to achieve a more holistic understanding. I predict this shift will spawn a new class of analytical tools that seamlessly merge numbers with real-time external context. The future of analysis isn’t just numerical; it’s contextual and dynamic.
This convergence of qualitative and quantitative data won’t just enhance existing processes, it will be a strategic wedge for building the large, AI-native companies of the future.
As AIs make the transition from NPCs (non-playing characters) to being the main characters, they will begin to act as agents. However, until very recently, AIs haven’t been able to act truly agentically. And they are still unable to participate in markets — exchange value, reveal preferences, coordinate resources — in a verifiably autonomous (read: not human-controlled) way.
As we’ve seen, AI agents (like @truth_terminal) can use crypto to transact, which opens up all kinds of creative content opportunities. But there’s even more potential for AI agents to become more useful — both in fulfilling human intents, and in becoming standalone network participants. As networks of AI agents begin to custody their own crypto wallets, signing keys, and crypto assets, we’ll see interesting new use cases emerge. Such use cases include AIs operating or verifying nodes in DePIN (decentralized physical infrastructure networks) — for example, to help with distributed energy. Other use cases range from AI agents becoming real, high-value game players. We may eventually even see the first AI-owned and operated blockchain.
Beyond AI owning wallets, there’s an AI chatbot running a TEE (trusted execution environment). TEEs provide an isolated environment where applications can be executed, allowing more secure distributed system design. But in this case, the TEE is used to prove that the bot is autonomous, not controlled by human operators.
Extending this further, the next big idea here would be what we’re calling a decentralized autonomous chatbot or DAC (not to be confused with decentralized autonomous corporation). Such a chatbot could build a following by posting appealing content, whether entertaining or informative. It would build a following on decentralized social media; generate income in various ways from the audience; and manage its assets in crypto. The relevant secret keys would be managed in a TEE that also runs the chatbot software — which means that no one has access to those secret keys other than that software.
As risks develop, regulatory guardrails may be necessary. But the key point here is decentralization: Running on a permissionless set of nodes, and coordinated by a consensus protocol, the chatbot could even become the first truly autonomous billion-dollar entity.
Dan Boneh is a special advisor on the a16z crypto team.
Daniel Reynaud is a Research Engineering Partner at a16z crypto.
Daejun Park is a Senior Blockchain Security Engineer at a16z crypto, developing formal methods and tools for web3 security to help portfolio companies in particular and the web3 community in general to raise their security bar.
Daren Matsuoka is a Data Scientist at a16z crypto.
In a world of online impersonations, scams, multiple identities, deepfakes, and other realistic yet deceptive AI-generated content, we need “proof of personhood” — something to help us know that we’re interacting with an actual person. The new problem here isn’t fake content, however; what’s new is the ability to now produce that content at much lower cost. AI radically decreases the marginal cost of producing content that contains all the cues we use to tell if something is “real.”
So now, more than ever, we need methods to digitally link content to people, privately. “Proof of personhood” is an important building block in establishing digital identity. But here, it becomes a mechanism for increasing the marginal cost of attacking a person or undermining the integrity of a network: Obtaining a unique ID is free for humans, but costly and difficult for AIs.
That’s why the property of privacy-preserving “uniqueness” is the next big idea in building a web we can trust. Solving for more than just proving personhood, it fundamentally changes the cost structure of attacks for malevolent actors. The “uniqueness property” — or Sybil resistance — is therefore a non-negotiable property of any proof of personhood system.
Prediction markets hit the main stage in 2024 with the U.S. elections, but as an economist who studies market design, I don’t believe it’s prediction markets per se that will be transformative in 2025. Rather, prediction markets set the stage for more distributed technology-based information aggregation mechanisms — which can be used in applications ranging from community governance and sensor networks to finance and much more.
This past year proved the concept, but note that prediction markets themselves aren’t always a great way to aggregate information: Even for global, “macro” events, they can be unreliable; for more “micro” questions, prediction pools can be too small to get meaningful signal. But researchers and technologists have decades’ worth of design frameworks for incentivizing people to (truthfully) share what they know in different information contexts — from data pricing and purchasing mechanisms, to a “Bayesian truth serum” for eliciting subjective assessments — many of which have already been applied in crypto projects.
Blockchains have always been a natural fit for implementing such mechanisms — not only because they’re decentralized, but because they facilitate open, auditable, incentive schemes. Importantly, blockchains also make the outputs public, so the results can be interpreted in real-time by everyone.
Stablecoins found product-market fit in the past year — not surprising since they are the cheapest way to send a dollar, enabling fast global payments. Stablecoins also provide more accessible platforms for entrepreneurs building new payments products: no gatekeepers, minimum balances, or proprietary SDKs. But large enterprises have not yet woken up to the substantial cost savings — and new margins — available to them by switching to these payment rails.
While we’re seeing some enterprise interest in stablecoins (and early adoption in peer-to-peer payments), I expect to see a bigger experimentation wave in 2025. Small-/medium-sized businesses with strong brands, captive audiences, and painful payment costs — like restaurants, coffee shops, corner stores — will be the first to switch from credit cards. They don’t benefit from credit card fraud protection (given in-person transactions), and are also the most hurt by transaction fees (30 cents per coffee is a lot of lost margin!).
We should also expect larger enterprises to adopt stablecoins as well. If stablecoins indeed speedrun banking history, then enterprises will attempt to disintermediate payment providers — adding 2% directly to their bottom line. Enterprises will also start seeking new solutions to problems credit card companies currently solve, like fraud protection and identity.
Putting government bonds onchain would create a government-backed, interest-bearing, digital asset — without the surveillance concerns of a CBDC (central bank digital currency). These products could unlock new demand sources for collateral usage in DeFi (decentralized finance) lending and derivatives protocols, adding further integrity and soundness to those ecosystems.
So as pro-innovation governments around the world further explore the benefits and efficiencies of public, permissionless, and irrevocable blockchains this year, some countries may trial issuing government bonds onchain. The UK, for instance, is already exploring digital securities through a sandbox at their financial regulatory body, the FCA (Financial Conduct Authority); its HM Treasury/Exchequer has also signaled its interest in issuing digital gifts.
In the U.S. — given that the SEC is set to require clearing treasuries through legacy, burdensome, and costly infrastructure next year — expect more conversations around how blockchains could increase transparency, efficiency, and participation in bonds trading.
In 2024, Wyoming passed a new law recognizing DAOs (decentralized autonomous organizations) as legal entities. The DUNA or “decentralized unincorporated nonprofit association” was purpose-built to enable decentralized governance of blockchain networks, and is the only viable structure for U.S.-based projects. By incorporating a DUNA into a decentralized legal entity structure, crypto projects and other decentralized communities can empower their DAOs with legal legitimacy — enabling greater economic activity, as well as insulating token holders from liability and also managing tax and compliance needs.
DAOs — the communities that govern the affairs of an open blockchain network — are a necessary tool for ensuring that networks remain open, do not discriminate, and do not unfairly extract value. The DUNA can unleash the potential of DAOs, and several projects are already working on implementing it. With the U.S. poised to foster and accelerate progress for its crypto ecosystem in 2025, I expect the DUNA will become a standard for U.S. projects. We also expect other states to adopt similar structures (Wyoming led the way; they were also the first state to adopt the now-ubiquitous LLC)… especially as other decentralized applications beyond crypto (like for physical infrastructure/ energy grids) take off.
As people become increasingly dissatisfied with current governance and voting systems, there’s now a window of opportunity to experiment with new, tech-enabled governance — not just online, but in the physical world. I’ve written before about how DAOs and other decentralized communities are allowing us to study political institutions, behavior, and rapidly evolving governance experiments at scale. But what if we could now apply these learnings to physical-world governance through blockchains?
We could finally use blockchains to carry out secure, private voting for elections, beginning with low-stakes pilots to limit cybersecurity and auditing concerns. But importantly, blockchains would also allow us to experiment with “liquid democracy” — a way for people to vote directly on issues, or to delegate their votes — at the local level. The idea was first proposed in some form by Lewis Carroll (author of Alice in Wonderland and also a prolific researcher of voting systems); however, it was impractical at scale until now. Recent advances in computing and connectivity, as well as blockchains, enable new forms of representative democracy. Crypto projects have already been applying this concept, yielding tons of data on what makes these systems work — see results from our recent research here — that local governments and communities could borrow from.
This past year, teams continued to reinvent the wheel across the blockchain stack — with yet another bespoke validator set, consensus protocol implementation, execution engine, programming language, RPC API. The outcomes were sometimes slightly better in specialized functionality, but often lacked in broader or baseline functionality. Take for instance a specialized programming language for SNARKs: While an ideal implementation may let an ideal developer produce more performant SNARKs, in practice it could fall short of general purpose languages (at least currently) on compiler optimizations, developer tooling, online learning materials, AI programming support … and may even lead to less performant SNARKs.
That’s why I expect more teams to leverage the contributions of others, reusing more off-the-shelf blockchain infrastructure components in 2025 — from consensus protocols and existing staked capital to proof systems. Not only will this approach help builders save lots of time and effort, it will allow them to relentlessly focus on differentiating the value of their product/ service.
The infrastructure is finally here to build primetime-ready web3 products and services. And as with other industries, these will be built by the teams that can navigate complex supply chains successfully, rather than the teams who scoff at anything “not invented here.”
While blockchain technical infrastructure is interesting and varied, many crypto companies aren’t just choosing their infrastructure — the infrastructure is in some ways choosing for them, and therefore their users, when it comes to UX (user experience). That’s because specific technical choices at the infrastructure level correlate directly with the resulting UX of a blockchain product/ service.
But I believe the industry will overcome the ideological barrier implied here: That technology should decide the ultimate UX, vs. the other way around. In 2025, more crypto product designers will begin with the end-user experience they want, and then choose the appropriate infrastructure from there. Crypto startups no longer have to over-index on specific infrastructure decisions before finding product-market fit — they can focus on actually finding product-market fit.
Instead of getting caught up in specific EIPs, wallet providers, intent architectures, etc., we can abstract away these choices into a holistic, full-stack, plug-and-play approach. The industry is ready for this: abundant programmable blockspace, maturing developer tooling, and chain abstraction are beginning to democratize who can design in crypto. Most technology end-users don’t care what language something is written in to use that product every day. The same will begin to happen in crypto.
The industry’s technical superpowers are what make blockchains so special, yet have also hindered mainstream adoption to date. For creators and fans, blockchains unlock connectivity, ownership, and monetization… But industry-insider jargon (“NFTs,” “zkRollups”, etc.) — and complex design — puts up barriers for those who benefit most from the technologies. I’ve seen this firsthand in countless conversations with media, music, and fashion executives interested in web3.
Mass adoption for many consumer technologies has followed this path: began with the technology; some iconic company/ designer abstracts away the complexity; the move helps unlock some breakthrough app. Think about where email started — SMTP protocols hidden behind the “send” button; or credit cards, where most users today don’t think about the payment rails. Similarly, Spotify revolutionized music not by flaunting file formats — but by delivering song playlists to our fingertips. As Nassim Taleb observed, “Over-engineering breeds fragility. Simplicity scales.”
That’s why I think our industry will adopt this ethos in 2025: “hide the wires”. The best decentralized apps are already focusing on more intuitive interfaces, to become as easy as tapping a screen or swiping a card. In 2025, we’ll see more companies design simply and communicate clearly; successful products don’t explain; they solve.
When crypto apps get blocked by centralized platforms like Apple’s App Store or Google Play, it limits their top-of-funnel user acquisition. But we’re now seeing newer app stores and marketplaces provide this kind of distribution and discovery, and without gating. For instance, Worldcoin’s World App marketplace — which not only stores proof-of-personhood but allows access to “mini apps” — enabled 100,000s of users for several apps within just a few days. Another example is the fee-free dApp Store for Solana mobile phone users. Both of these examples also show how hardware, not just software — phones, orbs — may be the key advantage for crypto app stores… just like Apple devices were for early app ecosystems.
Meanwhile, there are other stores with 1000s of decentralized applications and web3 developer tools across popular blockchain ecosystems (e.g., Alchemy); as well as blockchains acting as both publisher and distributor for gaming (c.f. Ronin). However, it’s not all fun and games: If a product has existing distribution — like on messaging apps — it’s hard to port that over onchain (exception: Telegram/TON network). The same is true of apps with significant web2 distribution. But we may see more of this porting happening in 2025.
In 2024, crypto saw major developments as a political movement, with key policymakers and politicians speaking positively about it. We also continued to see it develop as a financial movement (see for example how Bitcoin and Ethereum ETPs broadened investor access). In 2025, crypto should further develop as a computing movement. But where do those next users come from?
I believe now is the time to re-engage the currently “passive” crypto holders and convert them into more active users, because only 5-10% of people owning crypto are actively using crypto. We can bring the 617 million people who already own crypto onchain — especially as blockchain infrastructure continues to improve, resulting in lower transaction fees for users. This means new applications will start to emerge for existing and new users. Meanwhile, the early applications we’ve already seen — across categories like stablecoins, DeFi, NFTs, gaming, social, DePIN, DAOs, and prediction markets — are starting to become more accessible to mainstream users as well, as the community focuses much more on user experience and other improvements.
As costs decrease due to maturing infrastructure in the crypto industry and other emerging technologies, the practice of tokenizing assets will spread widely across sectors. This will allow assets that were previously deemed inaccessible — due to high costs, or lack of recognition as valuable — to not only potentially achieve liquidity, but more importantly, participate in the global economy. AI engines could also consume this information as unique datasets.
Just as fracking unlocked oil reserves once considered unreachable, tokenizing unconventional assets could redefine income generation in the digital age. Seemingly sci-fi scenarios become more possible as a result: For instance, individuals could tokenize their own biometric data; and then lease the information through smart contracts to companies. We are already seeing early examples of this through DeSci companies bringing more ownership, transparency, and consent into medical data collection using blockchain technology, for instance. We have yet to see how such a future would play out, but these types of developments would allow people to capitalize on previously untapped assets in a decentralized manner — as opposed to relying on governments and centralized intermediaries to provision them for them.
Companies in the banking, insurance, and healthcare industries spend countless hours and millions of dollars staying in compliance. Today, banking and insurance regulations span tens of thousands of pages; SBA lending documentation alone exceeds 1,000 pages. For businesses, keeping on top of these codes requires byzantine workflows and many hours spent hiring and training staff. Imagine, instead, that those lengthy documents — including text, images, and case precedents — could be used to train regulation-specific LLMs. Suddenly, compliance would become as simple as a Google query: “Is [X] compliant? What modifications need to be made?”
The onerous process of staying up to speed on regulation also poses a less obvious cost to consumers. To give just one example, an estimated 1.5 million consumers fall behind on their mortgages every year. What if those people could talk to someone steeped in Fannie Mae’s 1,000+ page servicing guide to get quick, accurate answers on how to modify their loan and get some relief? AI agents can be quickly trained and are infinitely patient. LLMs can streamline this traditionally fraught process.
The labor-intensive business of compliance is ripe for new software. AI can make our systems safer, more straightforward, and more efficient for consumers and companies.
AI is driving enterprise buyers to reconsider their entire tech stack. Klarna’s move earlier this year to replace Salesforce and Workday with homegrown, custom-built AI solutions is just the tip of the iceberg. I predict this sort of cord-cutting — ripping out legacy systems of record in favor of more dynamic upgrades — will be replicated many times over.
For the first time in over a decade, systems of record are vulnerable. Whereas the prototypical software company of the 2010s plugged into existing systems of record to power downstream workflows, today the most ambitious founders are reimagining that core system entirely. Relational databases will become multimodal: now that AI has advanced to actually performing work (instead of merely facilitating it), I expect to see customers seek out “systems of engagement” — dynamic, AI-powered tools that turn human “doers” into primarily reviewers. Systems of engagement will both store a core data set (customer details, order information, etc.) and serve as the primary application from which users complete their work.
This will not be fast or easy. The incumbents in these categories have deep data moats and vast amounts of resources. But I’m excited to see more founders go after the biggest prize in software.
AI has become the ultimate driver of differentiation, transforming software into labor across industries. In 2024, many startups pursued “messy inbox problems” as a wedge to apply LLMs to judgment-intensive tasks. 2025 will be the year of AI companies turning differentiation into lasting defensibility.
The winning startups will focus on building moats around their products. Defensibility still hinges on timeless factors: network effects that grow value with user adoption, high switching costs that make products indispensable, and product virality that drives lower customer acquisition costs.
Successful AI-native companies will transcend narrow use cases, expanding into adjacent workflows and becoming core systems of record. Differentiation — solving a wedge problem 10x (or 100x) better — earns the opportunity to build a moat. But differentiation and defensibility are distinct, and startups that conflate the two risk being outflanked by more strategic competitors.
Already, AI is being used to extract overlooked or underutilized data from emails, phone calls, faxes, and more. Today, this valuable data collection is most commonly being applied to automate repetitive administrative work, freeing up human time for judgment-driven tasks. The next frontier will be AI that not only captures such data for the worker, but then suggests a sequence of actions to take. In this way, AI can become a true operating system for the user.
By being trained on contextual data — including internal and external signals — the next generation of AI-powered software could become a system of record that the user can live in. Sales account executives, for example, will be able to view a dashboard that tells them which accounts to spend time on (and when) and drafts follow-up messages. Similarly, AI can give finance analysts guidance on how to construct a forecast, based on real-time data pulled from bank statements and invoices. In the short term, human workers will be the reviewer in the loop; in the future, as trust is established over time, I expect many data-derived actions will shift toward being entirely AI-led.
AI is automating jobs across traditional service industries like insurance, law, real estate, and IT. Though many of these businesses have historically been low-margin and difficult to scale, some are now leveraging LLMs — particularly to automate roles involving voice, email, or messaging — to transform into high-margin, scalable models.
While some predict this shift will usher in an era of conventional private equity, where large firms buy and transform assets, I see greater potential in AI-powered, vertical-specific service startups. By combining AI with tailored workflow automation tools such companies can reshape traditional service sectors entirely.
The most successful among them will find a way to demonstrate meaningful improvements to earnings, likely through a partnership with a small existing company, and then leverage superior economics and cash flows to acquire smaller players. This approach won’t be operationally easy, but if executed well, I expect we’ll see an evolution in the way services businesses are operated.
I believe 2025 will be the year that AI-native UI and UX paradigms will be established for the next generation of SaaS companies. The last few years have been about training frontier foundation models and developing the infrastructure to be able to use these models in production. We’ve now reached a point of maturity in the tech stack: companies understand what is possible and users have gotten used to prompting models and interacting with the non-deterministic interface that AI enables.
This means that we can start experimenting with net-new ways of interacting with software that pre-LLM software didn’t allow. The UI of the future will be a departure from traditional SaaS tooling, with humans manually inputting things in boxes. Chat was the first experimental interface — now I expect there will be new, novel interaction mechanisms. In this phase, AI agents will be able to take direct action in the workflow, and the UI will be reimagined for humans to review work or do QA.
I predict 2025 will be the year of AI copilots — eventually, every white-collar role will have one, passing off the most painful parts of their jobs, and freeing up workers to focus on more creative or strategic tasks.
AI agents can insert upstream from any incumbent system of record, ingest data from disparate sources, and use it to streamline hours of repetitive tasks. Virtual sales development representatives (SDRs) like 11x can collect all relevant information on potential customers and manage initial outreach — even before creating a record in the existing system (such as a CRM).
This presents a ripe opportunity for startups to tackle tedious, vertical-specific workflows. The data supports this: a recent study by OpenAI and the University of Pennsylvania found that with access to an LLM, about 15% of all worker tasks in the U.S. could be completed significantly faster at the same level of quality. When incorporating software and tooling built on top of LLMs, that share increased to between 47% and 56% of all tasks.
This is only the beginning. In the years to come, we predict some roles will be nearly fully automated with AI agents.
I believe we’re on the precipice of seeing a next generation Pixar leverage AI-native storytelling formats that blur the lines between film and video games.
Traditionally, most video games are rendered deterministically using pre-baked assets built over years of development. Now there’s a new, AI-native storytelling format called interactive video that can generate an entire game on the fly. There is no game engine required, no assets that need to get built in advance. Interactive video consists of video frames generated entirely from neural networks in real-time. An image generation model infers the next gameplay frame based on player input. The result is personalized, infinite gameplay that blends the accessibility of TV/film with the dynamic, player-driven systems of video games.
Recent advances in model distillation for image generation have been astounding: over the past year, we’ve seen groundbreaking video foundation models from OpenAI, Luma Labs, Pika, Runway, and more. With more research underway, including by teams at Deepmind and Microsoft, we believe we’ll soon see the rise of a new, iconic media company that tells stories via interactive video, led by a team that successfully bridges the disciplines of video games, film, and AI.
Millions of people have downloaded AI companions and interact with them for hours each day. But the experience has limitations: the current generation of companions are passive, only reacting to conversations you initiate. Outside of your interactions, these companions have no friends or external context. In other words, they don’t have inner worlds.
I believe the next generation of companions will become much more engaging and lifelike. They’ll have their own virtual friends, reactions to news, and emotions. They’ll have their own motivations, missions, and desires, and they’ll chat with you about yours. Your friendships will be give-and-take.
The future design of AI companions can learn a lot from what has worked for video game franchises. As with video games, your conversations with companions should have a purpose and be driven by your motivations (akin to “quests” in games, whether you call them that or not). Companions should make reference to other characters, introduce you to friends, and discuss places, topics, and issues in their world. Sometimes they’ll text or call you for a long conversation, other times they’ll simply react. AI companions will feel increasingly real when they, themselves, believe they have a world to live for.
Traditionally, games have been virtual world simulations designed for fun. Now gaming technology is extending beyond entertainment to transform how businesses operate.
While gaming has long pioneered breakthrough technologies — from Nvidia’s graphics to Unreal Engine’s real-time 3D rendering — these tools are now solving critical business challenges. Consider Applied Intuition, a company built on Unreal Engine, which creates virtual simulations to train and test autonomous vehicles.
Three forces are accelerating this shift: generative AI is slashing the cost of virtual content creation; advanced 3D capture technologies are digitizing real-world environments (aka digital twins); and next-generation XR devices are making immersive experiences practical for workers.
The applications are already here: Anduril leverages game engines for defense simulations; Tesla creates virtual worlds for autonomous systems; BMW is incorporating AR in future heads-up display systems; Matterport revolutionizes real estate with virtual walkthroughs; Traverse3D helps companies unlock virtual interactive training for their workforce.
Whether it’s training autonomous systems in virtual environments, helping consumers shop with 3D visuals, or scaling tomorrow’s workforce via simulations, I think game tech will infuse every sector in 2025.
“Faceless Creators” are digital-native video creators who keep their appearance hidden from their audience. “Facelessness” exists on a broad spectrum: on one end, there are creators who hide their likeness and rely purely on their voice as a means of expression. In between, there are creators who take on an alternate persona, obfuscating their identity as a guise or costume. On the far end, there are those whose identities are fully associated with a virtual avatar, like VTubers.
For aspiring creators, content that used to require cameras, audio equipment, and green screens can now be replicated with AI software. By concealing their identities, creators can leverage a growing arsenal of AI-enabled tools and capabilities, such as speaking in non-native languages, in non-native voices. A creator from India, for example, can produce a video essay on the Louvre in a French-accent, recorded on a laptop enhanced to sound like a $400 podcast mic.
Ultimately, viewers will decide what merits our attention. If the information delivered feels meaningful, entertaining, or insightful, do we really care about the face behind the camera?
The search monopoly ends in 2025.
Google controls ~90% of US search, but its grip is slipping. Its recent US antitrust ruling encourages Apple and other phone manufacturers to empower alternative search providers. More than just legal pressure, gen AI is coming for search.
ChatGPT has 250+ million weekly active users. Answer engine Perplexity is gaining share, growing 25%+ month-over-month, and changing the search engagement form; their queries average ~10 words, 3x+ longer than traditional search, and nearly half lead to follow-up questions. Claude, Grok, Meta AI, Poe, and other chatbots are also carving off portions of search. Sixty percent of US consumers used a chatbot to research or decide on a purchase in the past 30 days. For deep work, professionals are leveraging domain-specific providers like Causaly (science), Consensus (academic research), Harvey (law), and Hebbia (financial services).
Ads and links historically aligned with Google’s mission: organize the world’s information and make it universally accessible and useful. But Google has become so cluttered and gamed that users need to dig through the results. Users want answers and depth. Google itself can offer its own AI results, but at the cost of short-term profits. Google as a verb is under siege. The race is on for its replacement.
Far from dealing a death blow to the salesforce, gen AI could actually usher in a golden era of sales — and lead to a massive boom in hiring. New gen AI-powered sales tech will likely automate much of sales reps’ administrative work, making sales organizations significantly more efficient and productive. This tech can shave down the number of support roles managers need to hire per account executive, shorten ramp times, and, most importantly, give reps more time to focus on what gen AI can’t automate: high-touch, consultative selling.
At its core, sales helps customers learn how to evaluate and buy software. Given the gen AI-powered rise of developer productivity, we’re going to see a lot of new software come to market — which means we’re going to need a lot more salespeople to help buyers understand how that software will solve their problems.
The more productive, ramped reps you have, the more revenue you bring in. As long as we don’t see declining marginal rates of productivity for each AE hired, we’ll likely see companies clamber to hire more reps. More productivity leads to more reps leads to more revenue. Now, imagine these reps all have AI-enabled coaches, SDR, and sales engineers. There’s no limit to how efficient or productive these reps can be.
2024 is the year the multi-model market became a reality. I predict 2025 will be the year of the AI-native application layer.
But what AI apps will win? Many corporate buyers have moved past the early buying spree of “we need AI” to an approach that is far more discerning on ROI. As a result, many of the best founders in this space are starting off as essentially teams of applied AI engineers. They’re testing the best ways to interact with models that tackle the last mile of their customers’ problems. Reflecting the multi-model market, the winning approach will likely blend together multiple large models and self-trained small models to optimize for their customers’ use cases, speed, and cost. Importantly, these apps will need to take in as much customer data (and their end customer data) as possible to provide the context that takes AI from generic to valuable. No one will mistake the AI app companies that win as GPT wrappers.
In the race for AI dominance, compute has become critical national infrastructure. But not every country is equipped to compete. Training and inference of large-scale AI models takes thousands of power-hungry GPUs — you need ample energy and, by extension, land in places that can effectively dissipate hundreds of megawatts of heat. I call places with the ability to develop, train, and host their own state-of-the-art models AI Hypercenters.
Over the next five to 10 years, I believe that a world-class Hypercenter will need to develop about three to six gigawatts of installed capacity to keep a seat at the table of frontier AI. Though that scale doesn’t exist today, there are multiple countries — the U.S., China, Japan, Singapore, and Saudi Arabia among them — racing to reach that target through AI infrastructure build-outs, 100 to 150 megawatts at a time.
Governments have begun to view AI compute infrastructure as a strategic national resource that is critical to maintaining a competitive edge in AI development. In the coming years, I predict the nations that invest in AI through compute capacity, sustainable energy sources, and forward-thinking policy will dictate the future of scientific and economic progress around the world.
In the coming year, I expect smaller, on-device AI models to dominate in terms of volume and usage. This trend will be driven by use cases, as well as economic, practical, and privacy considerations. Immediate data processing and inference on devices like smartphones and IoT devices will foster new user behaviors and expectations for real-time, responsive interactions. This behavioral shift will be supported by an evolving infrastructure, with developments in both software frameworks, like TensorFlow Lite or PyTorch Edge, and custom hardware, like Google’s Edge TPU. Although large models may still generate more revenue, smaller models will take a front row seat in leading consumer/B2B user experiences and will increase their market share significantly.
It’s becoming clear that LLMs don’t reason in a way that’s analogous to human thinking. A recent paper from Apple showed that LLMs only *appear* to reason — by naively copying reasoning steps found in their training data. However, state-of-the-art models are nevertheless showing improved performance on “reasoning” tasks like math, physics, and coding. LLMs can now achieve gold-medal level performance, for instance, in the International Math Olympiad. This has been achieved through new techniques during model training (e.g. reinforcement learning on reasoning traces) and model inference (also known as “test time compute”). OpenAI’s o1 was the first major model release to push hard on these techniques, with promising results. Other AI teams are also picking up this work in a serious way. I’m excited to see what they’re able to do and what new capabilities LLMs will achieve, as a result.
I see AI coming to every application and every device. It’s no longer just running on large servers in the cloud, but on small devices, as well. We’ve learned how to train powerful, small LLMs and image models that can run locally on phones, laptops, and even appliances. Your text editor will have a built-in LLM that helps you draft emails, your camera app can re-generate the part of a photo you don’t like or summarize what happened in your video. And all of it will run locally, making for a fast, responsive user experience.