Weekly Notes: OpenSearch and Speedy Startups

Reflecting on billion-scale vector search infrastructure, AI agents, and how company building may be changing.

This week was about exploring two topics: first, what infrastructure changes it actually takes to improve semantic search at scale, and second, how modern startups are using agentic AI to punch above their weight and rival much larger companies.

Powering Billion-Scale Vector Search with OpenSearch

Article from Uber

  • TL;DR: Uber used Amazon OpenSearch to support vector search across billion-scale items. What stood out was not just the scale, but the number of practical tuning levers available, including better vector search algorithms and GPU acceleration.
  • Why it mattered: This was a useful reminder that production vector search is an infrastructure problem as much as a model problem. Relevance matters, but so do latency, cost, and the ability to tune the system without rebuilding everything from scratch.
  • My take: I liked seeing the logic behind moving from classic keyword search on Apache Lucene toward semantic search across Uber’s family of apps. What I still wanted, though, was a clearer comparison against the alternatives. The OpenSearch choice is interesting, but the trade-offs would have been even more useful than the final answer.
  • Practical takeaway: I want to pressure-test whether this same setup can also support RAG and agent memory, instead of treating those as separate systems by default.

The New Way To Build A Startup

Video on YouTube

  • TL;DR: The argument is that “20x companies” will automate most internal functions with AI agents, letting very small teams punch far above their weight. The two models that stood out were one AI super-employee handling many tasks, or one agent per employee built from that person’s workflows and documentation.
  • Why it mattered: This feels like more than a productivity upgrade. It reads like a rewrite of the startup playbook, where leverage delays hiring, lowers payroll, and lets a company stay small and cohesive for longer.
  • My take: I think this shift is real, and I think a lot of larger companies are underestimating how fast young startups can use it to move past them. At the same time, the labor side of this is still deeply unresolved. If employee data is used to train systems that can replace parts of their work, that changes the meaning of career growth in a pretty serious way.
  • Practical takeaway: Any team going deep on AI agents should think early about the employee and career implications, not just the automation upside.
Updated Apr 08, 2026 Back to top