← Back to blog

The GPT-3.5 to GPT-4 Inflection Point

Why the November 2022 and March 2023 moments reset product expectations, and how later jumps across Claude, Gemini, and DeepSeek repeated the same pattern.

  • GPT-3.5
  • GPT-4
  • AI History
  • Claude
  • Gemini
  • DeepSeek
  • Product Strategy
The GPT-3.5 to GPT-4 Inflection Point

The public AI timeline did not move in smooth increments. It moved in jumps, and GPT-3.5 to GPT-4 was the first mainstream one many builders felt directly.

Two anchor dates still matter:

  • November 30, 2022: ChatGPT shipped on GPT-3.5.
  • March 14, 2023: GPT-4 launched.

But they were the start of a broader pattern, not the end of it.

What GPT-3.5 changed

GPT-3.5 made language interfaces mainstream overnight. The key unlock was not perfect answers. It was conversational usability at scale.

For product builders, that changed three assumptions:

  1. Users would tolerate probabilistic outputs if the UX was fast and legible.
  2. Writing quality became a core product surface, not just a model artifact.
  3. “Good enough” assistants could create daily habits even with visible flaws.

What GPT-4 changed

GPT-4 raised the floor for reasoning-heavy tasks and long-context synthesis. This is where many teams learned a hard lesson: upgrading model quality does not fix weak product design.

The teams that won were the ones that paired better models with:

  • tighter prompts,
  • explicit response formats,
  • guardrails and recovery flows,
  • and real usage instrumentation.

The pattern repeated beyond GPT

You can see the same cycle in later releases too:

  • Claude 3 / 3.5,
  • Gemini 1.5,
  • GPT-4o,
  • DeepSeek-V3 and R1,
  • and the next open-model waves after that.

Different model families, same product dynamic:

  1. Demo shock.
  2. Over-promising by product teams.
  3. Reliability pain in production.
  4. A new wave of disciplined builders who add constraints and eval loops.

The core mistake is always the same: treating model capability as a product strategy.

Practical takeaway when any model drops

When a new model drops, I run the same checklist before touching roadmap scope:

QuestionWhy it matters
Does this reduce user effort in an existing workflow?Prevents novelty-driven detours
What error class does it actually improve?Forces measurable claim
What fallback is now required?Better models can still fail badly
Does this unlock a simpler UI?UX simplification is often the real value

If we cannot answer these in one meeting, we do not expand scope.

Bottom line

GPT-3.5 and GPT-4 were not just model milestones. They reset user expectations for speed, fluency, and usefulness.

Later model families confirmed the same thing.

The lesson was never “just use the latest model.” The lesson is: model jumps reward teams that can redesign the whole product loop quickly.


Best,
Oli
April 18, 2024