Featured

From a 243-Line GPT to Healthcare AI: What Happens When You Train Your Own Model

Recently, AI researcher Andrej Karpathy released a tiny but powerful project: MicroGPT.

A full GPT-style model.

Written in roughly 243 lines of pure Python.

No deep learning frameworks.
No GPU dependencies.
No complex infrastructure.

Just the raw algorithm.

For many engineers, it became something like a “Hello World” moment for large language models.

Because it proves something surprising:

The core logic behind modern AI systems is not thousands of files of code.

It’s a small set of mathematical ideas repeated at scale.

And that realization often leads developers to experiment.

Which is exactly what happened next.

Image

Training a GPT-Style Model on Healthcare Data

Inspired by this minimalist approach, I tried something different.

Instead of training on names or generic text, I trained a tiny GPT-style model on structured medical data.

Specifically:

ICD-10 codes and their descriptions pulled from ICD10Data.com.

The dataset came from a multi-tab Google Sheet.

Each sheet contained fragments of structured healthcare data that needed to be merged programmatically.

Once cleaned and unified, the dataset looked like this:

97,633 merged rows
77-token vocabulary
4,928 parameters

And just like MicroGPT, the entire training process ran with:

No PyTorch
No TensorFlow
No Hugging Face Transformers

Just pure Python with a simple autograd implementation.


The Model Architecture

The architecture was intentionally minimal.

Embedding → Linear → Softmax

That’s it.

The model operated at the character level, which means it learned from individual characters rather than tokens.

There was also:

No attention mechanism.
No context memory.
No transformer layers.

In other words, this was a deliberately constrained model.

The goal wasn’t performance.

The goal was understanding the learning dynamics.


What the Model Learned

After training, the model began generating outputs that looked like ICD-style medical entries:

A01.03 | Typhoid pneumonia

Were they perfect?

Not even close.

The training loss hovered around 4.34, which is close to a random baseline.

But something interesting still happened.

Even this tiny 5K-parameter model began learning the statistical structure of the dataset.

It learned:

• The ICD formatting pattern
• Code-description relationships
• Character distribution in medical terminology

In other words, the model learned structure without understanding.

And that’s exactly how large language models start.


What This Reveals About Large Models

When people look at systems like GPT-4 or future models like GPT-5, they often assume something mysterious is happening.

But the underlying mechanism is not magic.

The difference between a tiny notebook experiment and a frontier model comes down to five factors:

Scale
Attention mechanisms
Training data
Optimization techniques
Compute power

The mathematics is largely the same.

The magnitude is dramatically different.


Why Training Even a Tiny Model Matters

Running experiments like this changes how you think about AI systems.

When you train even a microscopic model yourself, several insights become obvious:

Why loss behaves the way it does.

Why context windows matter.

Why architecture often matters more than dataset size.

Why character-level modeling struggles with structured reasoning.

Why prompting works only when a model has memory and context.

These lessons are hard to grasp when working exclusively with API-based models.

But they become very clear when you build systems from the ground up.


What This Experiment Proved

The project produced several practical insights for healthcare AI.

Structured medical data can be programmatically merged and ingested into language models.

Domain-specific token distributions are statistically learnable.

Even complex-looking formats like ICD codes behave like structured language systems.

Most importantly:

Model architecture often limits capability more than dataset size.


What Comes Next

The next iteration of this experiment will introduce several upgrades:

Token-level modeling instead of characters.

Transformer attention layers.

Context windows that allow the model to reason across multiple tokens.

Because once context enters the system, everything changes.

Language models stop predicting isolated characters and begin modeling relationships between concepts.


The Bigger Lesson for Healthcare AI

Much of the current conversation around healthcare AI focuses on using APIs.

Plugging models into workflows.

Calling inference endpoints.

But building meaningful AI systems in healthcare requires something deeper.

It requires understanding the mathematics well enough to adapt the models to domain-specific data.

Experiments like this demonstrate that the journey toward advanced healthcare AI doesn’t always begin with massive infrastructure.

Sometimes it starts with a small dataset.

A tiny model.

And a few hundred lines of code.

From there, everything else scales.

Featured

Why Companies Like Meesho Are Quietly Building AI Voice Infrastructure

At first glance, an AI voice role at a company like Meesho might look like just another AI hiring trend.

Another chatbot.
Another voice assistant.
Another automation experiment.

But when you break it down, it reveals something much bigger.

This isn’t about building a voice bot.

It’s about rebuilding the conversation layer of a company’s operations.


The Hidden Operational Reality

Consider how a large marketplace actually functions day to day.

Behind the product interface, there are thousands of operational flows happening constantly.

Orders drop off before checkout.
First-time buyers abandon carts.
Returning customers stop reordering.
Delivery agents need coordination.
Vendors are recruited and onboarded through third parties.
Customer feedback must be collected continuously.

Now ask a simple question:

How many of these processes involve someone calling someone else?

The answer is: a lot.

And human conversation at scale is expensive.

Large companies often maintain entire teams dedicated to:

Customer outreach
Order confirmations
Feedback calls
Vendor onboarding
Workforce coordination

Each conversation requires a person, time, and operational overhead.

Multiply that across millions of users and the cost structure becomes significant.


The Real System Being Built

What companies are now building isn’t simply voice automation.

It’s conversation infrastructure.

Imagine the system architecture behind an AI-driven operational layer.

1. Campaign Layer

Teams in sales, operations, or customer success define campaigns.

They decide:

Target audience
Campaign objective
Trigger conditions

For example:

Customers who abandoned checkout
Users who haven’t reordered in 30 days
New delivery partners awaiting verification


2. Audience Layer

The system identifies specific user segments.

Examples include:

First-time buyers
High-value customers
Dormant users
Potential delivery agents

Instead of manual outreach lists, the system dynamically generates call targets.


3. Agent Layer

Here the voice agent executes conversations.

The agent can ask questions like:

Why didn’t you complete your purchase?
Would you like to reorder the product?
Can we confirm your delivery slot?
Can you share feedback on your experience?

The conversation becomes structured data collection rather than informal interaction.


4. AI Layer

Under the hood, the technical stack looks something like this:

Speech-to-text converts voice to language.

Large language models process intent and reasoning.

Text-to-speech generates responses.

This loop enables dynamic conversations rather than rigid scripts.

Technologies like ChatGPT and other LLM-based systems are increasingly used as the reasoning layer in these architectures.


5. Insight and Workflow Layer

Once conversations are complete, the real value appears.

The system can:

Extract key metrics
Trigger follow-up actions
Send payment links
Escalate cases to human agents
Log structured customer insights

In effect, every conversation becomes a data pipeline for operational intelligence.


Why This Matters for Unit Economics

The biggest shift here isn’t technological.

It’s financial.

Imagine a company that previously required ten manual callers to handle outreach campaigns.

Now imagine that same workload being handled by one autonomous AI system.

If outreach volume doubles while labor costs shrink, something important happens.

Margins expand.

This isn’t simply automation.

It’s cost-structure redesign.

And companies investing heavily in these systems are often signaling something important:

They are aggressively optimizing operational efficiency.


A Larger Industry Pattern

This shift isn’t unique to one marketplace.

Across industries—marketplaces, fintech platforms, logistics companies, and gig networks—the same operational leakages exist.

Abandoned carts.
Incomplete onboarding.
Manual feedback collection.
Delivery workforce coordination.
Vendor recruitment pipelines.

Nearly all of these processes involve conversation.

Which means they are prime candidates for automation.


From AI Assistants to Operational Workforce

Much of the public discussion around AI still focuses on assistants.

Tools that help employees write emails.
Tools that summarize documents.
Tools that support productivity.

But what companies are increasingly building looks different.

The progression appears to be:

AI as assistant
AI as automation
AI as operational workforce

In this model, voice is simply the interface.

The real system is autonomous workflow execution powered by conversation.


The Strategic Signal

When companies begin investing in systems like:

AI voice infrastructure
Autonomous agents
Internal workflow builders

They are not merely adding features.

They are redesigning the operational backbone of the organization.

The goal isn’t necessarily replacing humans.

It’s determining which conversations must remain human—and which can become programmatic.

Companies that solve that problem effectively will operate:

Leaner
Faster
More margin-efficient

And in highly competitive industries like marketplaces, that operational advantage can be decisive.


Voice Is Just the Entry Point

For now, voice is the visible layer of this transformation.

But the deeper shift lies in autonomy.

As AI systems become more capable, they won’t simply conduct conversations.

They will execute workflows.

Schedule deliveries.
Collect payments.
Trigger logistics updates.
Coordinate service providers.

Voice may introduce the system.

But autonomous operations will define its long-term impact.

And the companies building this infrastructure today may quietly be redesigning how modern digital businesses run.

Featured

How Convin Built Its Contact Center LLM

Image

Convin took a bold step: instead of relying on GPT-style APIs, it built its own 7-billion-parameter LLM, trained on 200B tokens with deep coverage of Indic and code-mixed languages. The result is a system tailored for contact centers, where agents switch between Hindi, English, and regional dialects in the same sentence.

But the model itself is only the foundation.

The real story is how Convin packages this LLM into agents, campaigns, and workflows—and how it enforces cost control, guardrails, and ROI at scale.

1. One Brain, Many Hats

An LLM on its own is like a hyper-smart intern: lots of raw knowledge, little discipline. What makes it useful in production is wrapping the base model into specialized agents:

  • RefundBot: handles damaged-item replacements, cites policy docs, calls replacement APIs.
  • CollectionsBot: negotiates payments, checks balances, updates ledgers.
  • QA Agent: reviews call transcripts, scores compliance, generates coaching tips.

Each agent is the same underlying LLM, but with its own prompt pack, knowledge base slice, tool permissions, and guardrails. That’s what turns raw intelligence into reliable workflows.


2. Campaigns, Not Just Models

In practice, enterprises don’t say “our LLM is deployed.” They say:

  • “Campaign X is live on outbound dialer list Y.”
  • “Agent Z is assisting with inbound IVR menu 2.”

Campaigns are the unit of value: they define who the agent serves, what data it can access, and what KPI it’s optimizing. This framing makes it possible to measure ROI: AHT reduction, CSAT uplift, collection yield, QA coverage.

The LLM is just the engine — campaigns are the steering wheel.


3. Guardrails Are Non-Negotiable

Hallucinations that pass as “creative” in chatbots are failures in regulated industries. A contact-center agent making up refund policies is a liability.

Platforms enforce guardrails at multiple layers:

  • Must-cite policies: claims about price, eligibility, or policy must include a KB reference.
  • Refusal protocols: if the evidence isn’t there, the model escalates — not improvises.
  • PII filters: OTPs, CVVs, Aadhaar numbers are blocked before the model even sees them.

The lesson: guardrails aren’t an afterthought — they are the foundation of trust.


4. Cost Discipline Is a Feature, Not an Afterthought

Every extra token is money. At scale, cost creep is inevitable unless you design for efficiency.

Strategies observed:

  • Cache repetitive queries: “How do I refund?” doesn’t need to hit the LLM 100 times/day.
  • Classify before you generate: let lightweight classifiers handle the easy 90%, reserve the LLM for the hard 10%.
  • Quantize aggressively: INT8 or INT4 models keep latency low and costs predictable.
  • Async caching: pre-generate common responses in advance, or fail once gracefully and cache the answer for next time.

Lesson: optimizing cost is as important as optimizing accuracy.


5. Prompts Are the New Features

We used to engineer features in ML pipelines. Now, we engineer prompts.

Each agent ships with a “prompt pack”: system rules, tone, refusal conditions. These prompts are:

  • Versioned like code
  • A/B tested in campaigns
  • Auto-promoted when one variant outperforms others

In production, “prompt ops” feels remarkably similar to “feature engineering ops.”


6. Data Is the Real Moat

Anyone can download an open-source LLM. The differentiator is the data flywheel:

  • Train on 200B tokens of multilingual + codemix corpora.
  • Fine-tune on millions of real call transcripts (de-identified).
  • Continuously improve adapters from QA feedback loops.

The result: a smaller, domain-specific LLM can outperform bigger general-purpose models. The moat isn’t model size — it’s data relevance.


7. The Business Model Is Layered

What ties it all together is how the LLM is monetized. Contact-center deployments use a mix of:

  • SaaS per seat (agents, managers, auditors)
  • Usage-based (per call, per minute, per token)
  • Campaign-based (per outbound list)
  • Enterprise licensing (flat annual contracts with SLAs)
  • Outcome-based (shared upside on collections or QA cost savings)

The LLM → Agents → Campaigns → Business Model stack ensures that the technical investment translates into revenue and ROI.


Final Takeaway

Deploying an LLM isn’t just “train a model and ship it.”

It’s:

  • One brain → many agents
  • Agents → campaigns with KPIs
  • Guardrails → trust
  • Cost discipline → sustainability
  • Prompts → features
  • Data → moat

That’s how a contact-center LLM transforms from a lab demo into a production-grade system.

AI Pricing Isn’t a Feature — It’s a System

I came across a job post for an AI-driven pricing Product Manager at Lenskart.

At first glance, it looks like another “AI + product” role.

But instead of thinking cool job, a different question came to mind:

What are they actually hiring this person to build?

Because pricing at a company like Lenskart isn’t a spreadsheet problem.

It’s a systems problem.


The Hidden Complexity Behind Retail Pricing

Think about the number of inputs involved in something as simple as pricing a pair of glasses.

You’re dealing with:

• Hundreds — maybe thousands — of SKUs
• Inventory attributes like material, frame type, and seasonal trends
• City-level demand differences (Delhi ≠ Bangalore ≠ Chennai)
• Known vs unknown customers
• New buyers vs repeat buyers
• App behavior and browsing patterns
• Cart abandonment and drop-offs
• Purchase history
• Returns and replacements
• Post-purchase feedback
• Offline stores and online channels

Now imagine all of that data streaming in every single day.

Pricing is no longer a static decision.

It becomes a living system.


The Real Question: What Should This Product Cost Right Now?

In a dynamic retail environment, pricing models constantly ask:

What should this product cost
right now
for this customer
in this city
through this channel?

That’s fundamentally different from traditional retail pricing.

Instead of periodic updates, the system must continuously evaluate signals from across the business.

This is where AI becomes useful.

Not as a dashboard.

Not as a feature.

But as infrastructure.


Why Pricing Systems Are Hard to Build

Predicting demand is only one piece of the challenge.

The real difficulty lies in building a system capable of coordinating multiple data streams simultaneously.

A robust pricing engine needs to:

• Ingest data from inventory, behavior, and feedback systems
• Reconcile online and offline signals
• Update prices dynamically without creating customer confusion
• Scale across multiple cities and product categories
• Remain explainable when finance or operations ask why

That last point matters more than most people realize.

When revenue changes, leadership doesn’t just want predictions.

They want traceability.


The Hardest Part: Closing the Feedback Loop

Even the most sophisticated pricing model is useless if it cannot learn from outcomes.

Once a price changes, the system must quickly measure what happened.

Did conversions increase or drop?

Did return rates change?

Did customer perception shift?

Did sales simply move to another SKU?

Without fast feedback loops, the system isn’t optimizing.

It’s guessing faster.

And guessing at scale can be expensive.


What This Role Really Signals

Roles like this reveal something interesting about how companies are evolving.

Organizations aren’t just hiring PMs to “use AI.”

They are hiring people who can orchestrate data into decisions.

That means building systems where:

Signals flow continuously.

Models update intelligently.

Outcomes feed back into the next decision cycle.

Pricing may be the visible layer.

But underneath it sits one of the most complex real-time decision engines in retail.


The New Shape of Product Leadership

Modern product leadership is changing.

It’s no longer just about launching features.

The real challenge is designing systems that can:

Observe
Decide
Learn
Adapt

Over and over again.

In that world, AI isn’t the product.

Learning systems are.

In Healthcare, MVP Isn’t Enough. It Has to Be MCSP.

In Silicon Valley, a broken MVP is part of the culture.

Ship fast.
Test quickly.
Fix what breaks.

If the product crashes, it’s called learning.

But healthcare operates under a very different reality.

Because in medicine, a broken system isn’t just a bug.

It can be negligence.


The Problem With “Move Fast and Break Things” in Medicine

Startup culture celebrates the Minimum Viable Product (MVP).

The idea is simple:

Launch early.
Validate demand.
Improve iteratively.

This approach works well in many industries.

A social app can afford bugs.
An e-commerce feature can fail temporarily.
A dashboard can return the wrong chart.

But healthcare systems interact with something fundamentally different:

Human lives.

You cannot A/B test patient safety.

You cannot deploy flawed clinical reasoning and wait for user feedback to refine it.

And yet, many founders entering healthtech bring Silicon Valley’s product philosophy with them.

That’s where the tension begins.


The Dangerous Confusion

Many teams conflate two very different types of “broken.”

One is acceptable.

The other is not.

Clunky UI — Acceptable

Early products can have rough interfaces.

Buttons might be misplaced.
Workflows may be inefficient.
Design can evolve over time.

These are usability problems.

They frustrate users, but they don’t endanger patients.


Unverified Clinical Logic — Unacceptable

Clinical algorithms, diagnostic tools, and decision-support systems must meet a far higher standard.

If a model recommends the wrong medication…

If a triage system misses a critical symptom…

If a risk score misclassifies a patient…

The consequences are not inconvenience.

They are clinical harm.


Introducing MCSP: Minimum Clinically Safe Product

Instead of chasing MVPs, healthtech startups should adopt a different principle:

MCSP — Minimum Clinically Safe Product.

The idea reframes product development around patient safety.

An MCSP accepts certain limitations:

• The interface can be simple.
• The feature set can be narrow.
• The product can launch in a limited scope.

But one thing must be non-negotiable:

The clinical reasoning must be validated before it touches a patient.


What Clinical Safety Actually Requires

Building an MCSP means prioritizing validation over speed.

That often includes:

Clinical expert review
Retrospective validation on real datasets
Peer-reviewed evidence where applicable
Regulatory compliance when required
Clear guardrails for edge cases

These processes can feel slow compared to traditional startup cycles.

But they exist for a reason.

Medicine operates on evidence, not iteration speed.


The Trust Economy of Healthcare

Every healthcare product operates inside a fragile ecosystem of trust.

Patients trust clinicians.
Clinicians trust tools.
Hospitals trust vendors.

Once that trust is broken, it’s extremely difficult to rebuild.

A single poorly validated AI recommendation can damage credibility for an entire company.

Or worse, undermine trust in digital health more broadly.

That’s why healthcare is fundamentally different from other technology markets.

The currency isn’t growth.

It’s trust.


Speed Still Matters—But Not at the Cost of Safety

None of this means innovation should slow down.

Healthcare desperately needs better tools, better data systems, and smarter AI.

Speed still matters.

But speed should apply to:

Workflow improvements
User interface iteration
Operational efficiency

Not to clinical correctness.

Those two timelines must remain separate.


The New Standard for Healthtech

If AI and software are going to reshape medicine, the industry must adopt a new product philosophy.

Not MVP.

MCSP.

A system that may be small, imperfect, and evolving—but clinically sound.

Because in healthcare, shipping something broken isn’t a learning experience.

It’s a liability.

And trust is far too valuable to spend on a bad MVP.

Your AI Output Is a Mirror of Your Instruction Clarity

Most professionals are still frustrated with AI tools.

They say the answers are shallow.
Generic.
Unstructured.

But the real issue usually isn’t the model.

It’s the instruction architecture.

People still prompt AI like this:

“Help me analyze market expansion.”
“Write a Python script for a report.”
“Create a PRD for a recommendation engine.”
“Explain this dataset.”
“Find sentiment in this chat.”

Then they blame the AI when the response feels vague.

But there’s a fundamental shift happening in how effective users interact with systems like ChatGPT.

The shift is simple:

AI responds to architecture.


1. Specificity: Vague Inputs Create Vague Outputs

AI models interpret instructions literally.

When you provide vague prompts, the system has to guess what you actually want.

Imagine asking a new analyst on your team:

“Analyze market expansion.”

Without context, they might not know:

Which market
Which timeframe
Which metrics matter
Which format you expect

AI behaves the same way.

Clear prompts define:

Scope
Objective
Audience
Depth of analysis

Specificity transforms output quality immediately.


2. Structure: AI Follows the Framework You Give It

Strong prompts separate instructions into clear sections.

For example:

Context
Instructions
Constraints
Output format

Without structure, the model improvises.

Sometimes it guesses correctly.

Often it doesn’t.

By defining structure, you guide the reasoning process.

The AI doesn’t just generate text—it executes a workflow.


3. Examples: Show What Good Looks Like

Many people rely on vague instructions like:

“Make it better.”
“Be sharp.”
“Write like a strategist.”

But language models learn best from examples.

Instead of describing the desired style, show it.

Provide a short sample of the tone, format, or reasoning depth you want.

This technique—known as few-shot prompting—dramatically improves accuracy and consistency.


4. Force the Reasoning Process

Another powerful technique is explicitly instructing the model to reason step by step.

For example:

“Think through the problem step-by-step before producing the final answer.”

This approach activates what researchers call chain-of-thought reasoning.

It allows the model to break complex problems into smaller intermediate steps.

The result is often:

Higher accuracy
Better logic
More transparent reasoning


5. Constraints Eliminate Generic Output

Most AI responses default to safe, generic language.

You’ll often see:

Long disclaimers
Generic introductions
Fluffy explanations

Unless you tell the model otherwise.

Effective prompts define constraints like:

No disclaimers
No generic introductions
Focus on actionable insights
Use bullet points

Constraints narrow the solution space.

And that typically produces sharper results.


The Deeper Insight: AI Is a Cognitive Interface

When prompts are carefully designed, something interesting happens.

The AI stops behaving like a chatbot.

Instead, it starts functioning more like a professional collaborator.

Depending on how you structure the prompt, it can operate as:

A research analyst
A strategy consultant
A product architect
A systems thinker

The difference isn’t the model.

It’s the instruction design.


The Rise of Instruction Literacy

Throughout history, influential thinkers often relied on structured mental frameworks.

Economists have models.
Strategists have playbooks.
Engineers have architectures.

AI interaction is evolving in a similar way.

The most effective users don’t just ask questions.

They design thinking blueprints.

This skill is sometimes called prompt engineering, but the deeper concept is instruction literacy—the ability to translate human intent into structured instructions for intelligent systems.


A New Form of Executive Leverage

By 2026, AI capability is no longer the bottleneck.

The models are already powerful.

What’s still catching up is how people instruct them.

Professionals who learn to architect prompts gain a powerful advantage.

They can turn AI systems into tools for:

Strategic analysis
Product design
Market research
Operational planning

In that sense, prompt design isn’t just a technical skill.

It’s becoming a form of executive leverage.

Because the quality of your AI output is ultimately a reflection of the clarity of your thinking.

2026 VC Predictions Market Map

Source: Murph Dataset: 132 VC investors Date current as of: February 10, 2026

The map aggregates forward-looking commentary from 132 venture investors and classifies their expectations into nine thematic buckets. It distinguishes between:

  • 🟢 Core Bets (repeated emphasis / strong conviction)
  • 🟢 Secondary Signals (emerging or mentioned once)

Rather than being a ranking, it’s a conviction-density map showing where investor narrative momentum is concentrating.

Article content

I. The Structural Thesis for 2026

Across the 132 investors, one meta-pattern stands out:

2026 is viewed as the year AI shifts from models to systems, from experimentation to economics, and from tools to autonomous operators.

There is noticeably less focus on:

  • Foundation model novelty
  • Raw model performance benchmarks
  • GPU hype cycles

And significantly more focus on:

  • Agent behavior
  • Infrastructure durability
  • Defensibility and moats
  • Revenue quality
  • Liquidity recovery

The market has moved from “what’s possible?” to “what scales and compounds?”


II. The 9 Major Themes — Deep Breakdown


1️⃣ Agentic Systems (Most Concentrated Conviction)

This appears to be the strongest clustering theme.

What VCs Mean by “Agentic Systems”

  • Multi-step autonomous agents
  • AI replacing workflows, not assisting them
  • Agents operating across apps (cross-tool execution)
  • Long-horizon task completion
  • Persistent goal-oriented systems
  • AI employees vs AI copilots

Why This Is a Core Bet

Investors believe:

  • The real value capture moves from models → orchestration layers
  • Enterprises want outcomes, not model access
  • Agents shift AI from CapEx experimentation to OpEx substitution

There’s heavy emphasis on:

  • Vertical agents (legal, sales, finance, healthcare)
  • Multi-agent coordination
  • Evaluation tooling for agents
  • Reliability and control layers

Signal: The “AI wrapper fatigue” narrative is over. Structured agent workflows with execution power are viewed as defensible.


2️⃣ AI Infrastructure (Still Critical, but Evolving)

This remains one of the densest categories.

But tone has shifted from:

  • GPU arms race
  • Scaling laws obsession

To:

  • Cost compression
  • Inference optimization
  • Fine-tuned stack components
  • Open-source leverage

Focus areas include:

  • Inference efficiency
  • Orchestration layers
  • Model routing
  • Observability
  • Enterprise deployment infra
  • On-device models
  • Security layers

Infrastructure in 2026 = cost control + reliability + enterprise readiness


3️⃣ Multi-Modal & Reasoning Systems

Strong but less dominant than agents.

Investor focus:

  • Video-native models
  • Long-context reasoning
  • Tool usage integration
  • Structured output improvement
  • Planning-capable systems

Reasoning is viewed as:

  • Necessary for agents to succeed
  • A key differentiation axis
  • Critical for enterprise trust

However, this theme often appears as a supporting bet, not standalone.


4️⃣ Physical AI & Energy

A growing (but not universal) conviction cluster.

Subthemes:

  • Robotics + AI convergence
  • Industrial automation
  • AI-powered manufacturing
  • Autonomous mobility
  • AI energy demands
  • Nuclear / grid tech adjacency

Energy is increasingly discussed due to:

  • AI data center load
  • Geopolitical supply concerns
  • Infrastructure bottlenecks

This theme feels early but rising.


5️⃣ AI Economics & Pricing

This is where the maturity shift becomes obvious.

Investors are deeply focused on:

  • Usage-based pricing models
  • AI replacing labor
  • Margin structure of AI businesses
  • Token cost compression
  • AI-native SaaS revenue models
  • Revenue share or outcome pricing

2026 is viewed as:

“The year AI business models are tested.”

There is growing scrutiny of:

  • Gross margins
  • Defensibility beyond API access
  • CAC vs value capture

6️⃣ Data & Memory Moats

Defensibility narrative intensifies here.

VCs emphasize:

  • Proprietary data loops
  • Memory systems
  • Persistent user history
  • Workflow capture
  • Context accumulation

Key belief:

The only durable moat in a fast-improving model world is proprietary data and recurring interaction depth.

Memory (long-term context retention) is emerging as:

  • A product feature
  • A defensibility layer
  • A switching cost mechanism

7️⃣ Talent & Distribution

This theme speaks to market structure.

Recurring notes:

  • AI-native founders
  • Solo billion-dollar companies
  • AI-enhanced teams
  • Distribution-first startups
  • Brand leverage as moat

Investors are acutely aware:

  • AI reduces team size
  • Distribution matters more than code
  • Traditional SaaS playbooks may not apply

Distribution moats often appear alongside agent bets.


8️⃣ Fintech & Financial Rails

Moderate density cluster.

Includes:

  • AI in underwriting
  • AI lending infrastructure
  • Embedded finance
  • Cross-border rails
  • Stablecoin infra (often implied)

AI is seen as:

  • Risk underwriting enhancer
  • Fraud detection amplifier
  • Pricing engine

But fintech is not dominant in the 2026 prediction landscape — it’s selective and efficiency-driven.


9️⃣ Liquidity & IPO Market

One of the more interesting macro layers.

Investors are watching:

  • IPO window reopening
  • Exit liquidity timing
  • Secondaries
  • AI IPO wave

There’s cautious optimism:

  • 2026 may mark broader liquidity normalization
  • AI pure-play public listings expected

This is not a thematic investment category but a capital-cycle signal.


III. Cross-Cutting Meta-Trends

Across the grid, several structural transitions are visible.


1. Model Wars → System Wars

Less commentary on which foundation model wins. More focus on:

  • Orchestration
  • Application layers
  • Distribution

The center of gravity is shifting upward in the stack.


2. “AI as Tool” → “AI as Worker”

Language shifts from copilots to:

  • Autonomous agents
  • AI employees
  • Replacement economics
  • Revenue per AI headcount

This is a conceptual leap in investor framing.


3. Cost Compression as a Tailwind

Investors expect:

  • Token prices to fall
  • Inference to get cheaper
  • Open-source to expand

This unlocks:

  • Vertical AI startups
  • Lower barriers to entry
  • Margin expansion later in the cycle

4. Defensibility Anxiety

Implicit throughout the map:

“What actually builds a moat?”

Common answers:

  • Data loops
  • Embedded workflows
  • Memory systems
  • Distribution networks
  • Brand

Simple model wrappers are rarely presented as durable.


5. Enterprise Adoption > Consumer Hype

Most density clusters around:

  • Enterprise AI
  • Infrastructure
  • Vertical productivity

Consumer AI appears less frequently as a pure standalone opportunity.


IV. What’s De-Emphasized

Notably less present:

  • Web3 speculation
  • Metaverse narratives
  • Pure social platforms
  • Crypto-native hype cycles
  • Model pre-training companies
  • Zero-revenue AI companies

The market tone is pragmatic.


V. What 2026 Looks Like in Venture Psychology

The map suggests investors believe:

  1. AI is no longer experimental
  2. Product-market fit standards are rising
  3. Revenue quality matters again
  4. Capital efficiency matters again
  5. Exit cycles may restart

In short:

2026 = Execution Year

Not invention year. Not hype year. Execution, monetization, and consolidation.


VI. If You Compress It Into One Sentence

2026 will reward AI companies that:

  • Build autonomous systems
  • Own proprietary data
  • Monetize real workflows
  • Control cost structure
  • And operate with disciplined business models
Image

The 243-Line GPT: Why a Tiny Script Is Changing How People Understand AI

In the world of artificial intelligence, models are usually associated with enormous complexity.

Billions of parameters.
Distributed GPU clusters.
Massive training pipelines.

But recently, a remarkably small piece of code has been quietly reshaping how people understand large language models.

In a widely shared GitHub gist, AI researcher Andrej Karpathy released MicroGPT—a dependency-free implementation of a GPT-style transformer written in roughly 243 lines of pure Python.

No deep learning frameworks.

No optimized libraries.

Just the raw algorithm.

And that simplicity is exactly the point.


Stripping GPT Down to Its Core

Modern AI development is often built on powerful frameworks like PyTorch or TensorFlow.

These tools abstract away enormous amounts of complexity.

But abstraction can hide the underlying mechanics.

Karpathy’s MicroGPT does the opposite.

It reveals that the essential algorithmic structure of a transformer-based language model can be expressed with:

  • Basic Python lists
  • Scalar automatic differentiation
  • A few matrix-style operations
  • Attention and feedforward layers

Everything else—GPU acceleration, distributed training, memory optimization—is primarily about efficiency and scale, not the fundamental idea.

As the project description puts it:

“Everything else is just efficiency.”


What the Script Actually Does

Despite its small size, the MicroGPT implementation contains all the essential components of a transformer language model.

The script includes:

  • A tokenizer that converts characters into tokens
  • A simple automatic differentiation engine
  • Transformer attention layers
  • A feedforward neural network
  • The Adam optimization algorithm
  • A training loop and inference pipeline

In short, the code implements the entire learning pipeline of a GPT-style model.

It trains on a small dataset—often simple text like lists of names—and learns to generate new samples by predicting the next token in a sequence.

The output might look trivial: invented names or small text fragments.

But the mechanism behind it is the same one that powers systems like GPT-4.

The difference is scale.


Why This Matters for AI Education

Projects like MicroGPT reveal an important truth about modern AI:

The core mathematics behind large language models is not fundamentally mysterious.

It is built from relatively straightforward ideas:

Probability distributions over tokens
Gradient-based optimization
Matrix multiplications
Attention mechanisms

The real complexity of production models lies in engineering challenges.

Scaling models to billions of parameters.
Training on trillions of tokens.
Running efficiently across massive GPU clusters.

But the conceptual engine of these systems can still be understood in a few hundred lines of code.

For students and researchers, this dramatically lowers the barrier to learning how language models actually work.


The Explosion of MicroGPT Variants

Since its release, the project has sparked a wave of experimentation across the developer community.

Engineers have created alternative versions in multiple programming languages.

Implementations now exist in:

  • C++
  • Rust
  • Go
  • Julia
  • JavaScript
  • OCaml

Each variant explores a different aspect of optimization or architectural design.

Some aim to maximize performance.

Others focus on clarity and pedagogical value.

One optimized implementation reported speedups of more than 19,000× compared to the original Python version by rewriting the computation loops and exploiting CPU vectorization.

But even these optimized versions remain faithful to the same underlying algorithm.


The Importance of Iteration Speed

One of the most interesting insights emerging from these experiments involves training speed.

When training takes hours or days, researchers can test only a few hypotheses at a time.

But when training runs complete in seconds, the research process changes dramatically.

Developers can run:

Hundreds of hyperparameter experiments.
Rapid architecture tests.
Interactive training explorations.

This creates a feedback loop where faster experiments lead to deeper understanding.

As one contributor to the MicroGPT ecosystem described it:

“Speed doesn’t just make things faster. It changes what you can notice.”

In other words, iteration speed shapes discovery.


A “Hello World” Moment for Language Models

Some developers have compared MicroGPT to the classic “Hello World” program for large language models.

It provides a minimal but complete example of the technology.

Not optimized.

Not production-ready.

But conceptually complete.

Just as early computer scientists learned programming by examining small programs, modern AI researchers can now study language models in their most atomic form.

This transparency demystifies systems that are often perceived as opaque or inaccessible.


The Bigger Lesson

Projects like MicroGPT reveal something important about the trajectory of artificial intelligence.

Breakthroughs rarely come from a single algorithm or hardware innovation.

They emerge from compounding improvements across the entire research loop:

Better hardware enables more experiments.

More experiments produce deeper understanding.

That understanding leads to improved algorithms.

And those algorithms drive further hardware innovation.

The result is a cycle of accelerating progress.

A 243-line script may seem insignificant compared with billion-parameter models.

But it captures the core mechanism of modern AI in a form that anyone can read, modify, and experiment with.

And sometimes, understanding the engine matters more than building a bigger machine.

The Real Race in AI Agents Isn’t Models ; It’s Execution

A small announcement recently caught attention in the AI community.

TinyFish claims its web agent scored nearly 90% on the Online-Mind2Web benchmark—the highest score recorded so far.

The benchmark tests something deceptively simple:

Can an AI agent actually use the internet the way humans do?

Click buttons.
Fill forms.
Navigate pages.
Apply filters.
Complete multi-step workflows.

And if the agent makes the wrong move at any step, the task fails.

According to TinyFish, their agent achieved 89.97%, significantly outperforming several well-known systems:

  • Google Gemini computer-use agents — ~69%
  • OpenAI Operator — ~61%
  • Claude computer-use agents — ~56%

On paper, that gap looks dramatic.

But the interesting part isn’t the leaderboard.

It’s what the benchmark actually measures.


From Chatbots to Web Operators

For years, AI systems were evaluated primarily on language ability.

Could they generate text?
Write code?
Answer questions?

But real-world work rarely happens inside a chat window.

It happens inside software interfaces.

Logging into dashboards.
Submitting forms.
Searching internal systems.
Managing operational workflows.

The next generation of AI isn’t just generating answers.

It’s performing tasks inside digital environments.

And that’s exactly what web-agent benchmarks like Online-Mind2Web attempt to measure.


Why Benchmarks Like Mind2Web Matter

Traditional AI benchmarks test intelligence in isolation.

Web-agent benchmarks test interaction with messy real systems.

Agents must:

Understand a user’s objective.
Interpret a webpage interface.
Identify the correct button or field.
Execute the right sequence of steps.

Every mistake compounds.

Click the wrong menu?
The task fails.

Enter the wrong form value?
The workflow breaks.

Apply the wrong filter?
The agent ends up on the wrong page.

In other words, this benchmark measures something closer to real digital labor.


The Transparency Play

Another interesting aspect of the TinyFish announcement is transparency.

Instead of showing polished demos, the company published a spreadsheet with all 300 benchmark runs, including failures.

This approach reflects a growing trend in the AI ecosystem.

As benchmarks become more competitive, companies increasingly face criticism for:

Cherry-picked demos
Selective benchmarks
Carefully curated examples

Publishing full run logs helps address those concerns by making the evaluation process inspectable.

It’s not just about the final score.

It’s about showing how the system behaves step by step.


The Bigger Shift: Agents as Infrastructure

Announcements like this point toward a larger transformation in AI.

For the past few years, the focus has been on models.

Who has the most capable large language model?
Who can generate the best text or code?

But the next phase may revolve around something else entirely:

Agents that can execute workflows.

Instead of answering questions, AI systems will increasingly:

Book travel
Process invoices
Manage dashboards
Handle internal tools
Operate enterprise software

In this model, the AI becomes less like a chatbot and more like a digital operator.


Why This Changes the AI Competition

If agents become capable of reliably executing tasks on the web, the competitive landscape changes.

The most valuable systems won’t necessarily be the ones that write the most eloquent responses.

They will be the ones that can reliably complete workflows.

Logging into systems.
Navigating tools.
Submitting transactions.

This type of capability turns AI into something far more powerful than a conversational assistant.

It turns AI into a programmable workforce layer.


The Benchmark Arms Race

Benchmarks like Online-Mind2Web may soon become the equivalent of ImageNet for AI agents.

They provide a standardized way to compare systems across companies.

And they encourage rapid iteration.

But benchmarks are also just one part of the story.

Real-world deployments are far more chaotic than controlled evaluations.

Interfaces change.

Websites update.

Workflows evolve.

Which means the true challenge for AI agents isn’t just solving benchmarks.

It’s adapting to constantly shifting environments.


The Real Question

TinyFish’s result is impressive if it holds under scrutiny.

But the more important signal is what this type of progress represents.

AI systems are moving beyond generating language.

They are learning how to operate software environments.

And once agents can reliably interact with digital tools, the implications become much larger.

Because the internet itself becomes an execution environment for AI.

Not just a source of information.

But a place where autonomous systems can work.

And that shift may ultimately define the next phase of the AI era.

Why AI in Mammography Is Moving From Innovation to Standard of Care

A growing body of evidence suggests that artificial intelligence is poised to transform one of the most widely used cancer screening tools: mammography.

In a recent analysis, cardiologist and digital medicine researcher Eric Topol argued that AI analysis should be incorporated into every mammogram. His conclusion follows accumulating clinical evidence that AI can significantly improve cancer detection without increasing false positives.

At a time when breast cancer remains one of the most common cancers worldwide, the implications are profound.

Image

The Evidence: The MASAI Trial

Much of the recent momentum comes from results of the MASAI trial (Mammography Screening with Artificial Intelligence), one of the largest real-world evaluations of AI-assisted breast screening.

The study followed more than 105,000 women in Sweden, comparing two approaches:

  • Traditional radiologist interpretation
  • Radiologist interpretation supported by AI

The results were notable.

AI-assisted screening detected:

  • 29% more cancers overall
  • 24% more invasive cancers
  • No increase in false positives

At the same time, the use of AI reduced radiologist workload by roughly 30%, suggesting that the technology can improve both diagnostic accuracy and operational efficiency.

In a field where early detection dramatically affects survival rates, even incremental improvements can translate into thousands of lives saved.


Beyond Detection: Predicting Risk

The potential impact of AI mammography extends beyond simply finding tumors earlier.

Recent algorithms are beginning to extract additional clinical signals from the same imaging data.

Researchers have demonstrated that AI models can now:

  • Predict future breast cancer risk over the next five years
  • Detect breast arterial calcification, an indicator associated with cardiovascular disease
  • Identify subtle imaging patterns that may be invisible to human readers

In effect, a routine screening exam could evolve into a multidimensional risk assessment tool, offering insights into both cancer and heart health.

This capability represents a shift from reactive detection to predictive medicine.


The Adoption Gap

Despite promising results, widespread implementation of AI in mammography has been uneven—particularly in the United States.

Today, AI-assisted mammogram analysis is often available only through limited hospital networks and imaging centers.

In many cases, patients must pay an additional $40 to $120 out of pocket for AI analysis.

That creates a paradox.

If the technology demonstrably improves detection rates, why is it treated as an optional upgrade rather than a standard component of screening?

This question becomes even more urgent considering that the National Cancer Institute estimates that approximately 20% of breast cancers may be missed by traditional mammography alone.


A Policy and Incentive Problem

The barriers to adoption are rarely technological.

In many cases, they are economic and regulatory.

Healthcare systems often adopt new technologies slowly because reimbursement models, insurance coverage, and regulatory approvals lag behind scientific progress.

This can create situations where a clinically superior tool exists but is not widely deployed.

Critics argue that when evidence becomes sufficiently strong, AI-assisted screening should transition from an optional add-on to standard of care.

In practical terms, that would mean:

  • Integrating AI analysis into routine mammography workflows
  • Ensuring insurance coverage
  • Eliminating additional patient costs

Such steps would align financial incentives with clinical outcomes.


A Broader Trend in Medical AI

Mammography may represent one of the clearest early examples of how AI could reshape medical imaging.

Across radiology, AI tools are being developed to detect abnormalities in:

  • CT scans
  • MRI images
  • Chest X-rays
  • Pathology slides

In many cases, the most effective model is not AI replacing clinicians, but AI augmenting human decision-making.

Radiologists remain central to diagnosis and treatment decisions.

AI simply adds an additional layer of analysis—often identifying subtle signals that might otherwise be overlooked.


When Technology Becomes Infrastructure

Historically, certain medical innovations eventually become so clearly beneficial that they cease to be optional.

Sterile surgical tools.
Antibiotics.
Standard imaging protocols.

The argument now being made by researchers like Topol is that AI-assisted mammography may soon fall into that category.

If the technology consistently improves detection while reducing physician workload, the question shifts from whether to adopt it to why it has not yet become universal.


The Real Question Ahead

The science behind AI-assisted mammography is increasingly compelling.

But the real challenge may not be technological validation.

It may be the speed at which healthcare systems adapt to integrate new capabilities into everyday care.

Because when a technology exists that can catch cancer earlier and improve survival rates, the debate ultimately becomes simple:

Not whether it works.

But whether society chooses to use it at scale.

AI Doesn’t Make You an Expert ; But It Is Redefining What Expertise Means

A recent viral post made a blunt claim!

Using AI tools like ChatGPT to write marketing material, proposals, or slide decks does not suddenly make someone an expert.

The post goes further, arguing that people are attempting to fake expertise with AI—a practice it calls “charlatan nonsense.”

It’s a provocative statement.
And it touches a real tension emerging in the AI era.

But the deeper question isn’t whether AI creates fake experts.

The real question is:

What actually counts as expertise in a world where AI can generate knowledge on demand?


The Concern: AI as Shortcut to Authority

The concern behind the criticism is understandable.

For decades, expertise was earned through:

Years of practice
Industry experience
Repeated trial and error
Deep contextual understanding

Writing a strategy document or crafting a marketing plan required domain knowledge.

Today, tools like ChatGPT can generate:

Marketing copy
Business plans
Product positioning
Pitch decks
Technical explanations

In seconds.

This creates a fear that individuals might present AI-generated outputs as their own professional insight without truly understanding the subject.

And that concern isn’t entirely unfounded.

AI can produce convincing language.
But convincing language is not the same thing as expertise.


The Historical Pattern of “Tool Anxiety”

However, history shows that every major productivity tool initially triggers the same reaction.

When spreadsheets appeared, critics said financial professionals would lose real analytical skills.

When search engines emerged, educators worried that people would stop learning because answers were instantly available.

When calculators became common, many argued students would lose their mathematical ability.

In reality, something different happened.

These tools didn’t eliminate expertise.

They shifted the nature of expertise.

The professionals who thrived weren’t those who avoided the tools.

They were those who learned how to think better with them.


The Real Divide: Operators vs Thinkers

AI is creating a new distinction between two types of users.

The first group simply asks AI to generate content.

They copy outputs without questioning assumptions, verifying facts, or adapting the results.

The second group treats AI differently.

They use it as a thinking accelerator.

They refine prompts.
Challenge outputs.
Inject domain knowledge.
Iterate ideas quickly.

In this model, AI doesn’t replace expertise.

It amplifies the thinking of people who already understand the problem space.

The difference isn’t the tool.

It’s the judgment of the person using it.


The Rise of “AI-Assisted Expertise”

What we are likely witnessing is the emergence of a new professional skill:

AI-assisted expertise.

In this model, professionals combine:

Domain knowledge
Strategic thinking
AI-driven research and generation

The expert is no longer defined by their ability to manually produce every output.

Instead, they are defined by their ability to:

Frame the right questions
Evaluate generated information
Synthesize insights quickly

In other words, expertise shifts from production to interpretation and decision-making.


Why This Debate Matters

The debate about AI and expertise reflects a deeper transformation happening across knowledge industries.

Writers, consultants, analysts, designers, and strategists are all entering a world where AI can produce first drafts instantly.

The advantage will no longer belong to the person who can generate the most content.

It will belong to the person who can identify the most valuable ideas and refine them into strategy.

This is the difference between information generation and insight creation.

AI is exceptional at the first.

Humans remain essential for the second.


The Real Risk Isn’t AI

Ironically, the biggest risk isn’t that AI creates fake experts.

The real risk is something else entirely:

People who refuse to learn how to use it effectively.

Because the professionals who combine domain expertise with AI capability will move dramatically faster.

They will research faster.

Think faster.

Prototype ideas faster.

And iterate strategies faster.

In many industries, that speed advantage will define the next generation of leaders.


Expertise Is Not Disappearing ; It’s Evolving

AI does not magically turn someone into an expert.

But it also doesn’t invalidate expertise.

Instead, it is reshaping how expertise is expressed.

The most valuable professionals in the coming decade will not be those who reject AI.

Nor those who blindly rely on it.

They will be those who understand both the power and limitations of intelligent systems.

Because in the end, tools don’t define expertise.

Judgment does.