Moss (YC F25)’s cover photo
Moss (YC F25)

Moss (YC F25)

Technology, Information and Internet

San Francisco Bay Area, California 2,201 followers

Real-time Semantic Search for Conversational AI

About us

Moss is the real-time search runtime for conversational and multimodal AI. We enable voice agents, copilots, and chat interfaces to retrieve, and reason in under 10 milliseconds - making AI interactions feel truly natural. Moss runs natively across browsers, mobile devices, and servers with an optimized index built in Rust and WebAssembly. Drop-in SDKs for JavaScript and Python. No infrastructure to manage. Founded by Sri Raghu Malireddi (ex-Grammarly, Microsoft ML Lead) and Harsha Nalluru (ex-Microsoft, Azure SDK architect). Backed by Y Combinator (F25). Learn more at moss.dev

Website
https://moss.dev
Industry
Technology, Information and Internet
Company size
2-10 employees
Headquarters
San Francisco Bay Area, California
Type
Privately Held
Founded
2024
Specialties
Semantic Search, Real-time AI, Conversational AI, Voice AI, Vector Search, Edge AI, WebAssembly, and Developer Tools

Locations

Employees at Moss (YC F25)

Updates

  • Moss (YC F25) reposted this

    View profile for Sri Raghu Malireddi

    Moss (YC F25)3K followers

    I've been telling everyone Moss does sub-10ms retrieval. Time to prove it. So we did. We published full benchmark results on our GitHub repo this week. Open, reproducible, run them yourself. Here's why this mattered to me personally: 1/ I got tired of seeing infrastructure companies throw around performance numbers with no way to verify. I didn't want Moss to be one of those companies. If we're going to make the claim, the proof should be public. 2/ These aren't cherry-picked. We tested across real workloads, different dataset sizes, different query patterns. The results hold across the board. 3/ The number matters because of what it unlocks. Sub-10ms is the difference between a voice AI agent that feels instant and one that pauses. Between a search experience that's seamless and one where the user notices the wait. This was a bet we made early on. That speed would be the moat. That if we obsessed over single-digit millisecond retrieval, everything else would follow. The repo is open. The numbers speak for themselves. Link in comments.

    • No alternative text description for this image
  • 𝗕𝘂𝗶𝗹𝗱𝗲𝗿𝘀 𝗪𝗵𝗼 𝗦𝗵𝗶𝗽 𝗩𝗼𝗶𝗰𝗲 is back. Happy hour for Voice AI builders and founders in SF. Agents, pipelines, real-time infra, if you're building on the voice stack, come through. No panels. No pitch decks. Just good conversations with people actually shipping to production. 📍 San Francisco 📅 Thursday, 16th April · 5:30 – 8:30 PM → RSVP: https://luma.com/l2fb0s4s #VoiceAI #AIAgents #VoiceAgents #SFTech #Developers

  • 𝗧𝘆𝗽𝗲 𝗮𝘀 𝗳𝗮𝘀𝘁 𝗮𝘀 𝘆𝗼𝘂 𝗰𝗮𝗻. 𝗪𝗲'𝗹𝗹 𝗸𝗲𝗲𝗽 𝘂𝗽. We just shipped Moss semantic search plugin for VitePress docs. Runs entirely in the browser. Sub-10ms. Your user can type "how do I authenticate?" and gets the right answer, even if no page contains those exact words. Nothing sits between the keystroke and the answer. One plugin. Five minutes to set up. Try Demo: https://lnkd.in/gHuzTetJ

  • "Is this video edited, or is this real?" That's what a customer asked after we sent them a demo video of a voice AI agent we built for their use case. Here's the backstory. They came to us because their voice agent was lagging in retrieval-heavy conversations. They didn't share exact specs. Just said it was slow and the experience was falling apart. Instead of debugging their setup, I told the team, let's just build one from scratch with Moss underneath. Sub-10ms retrieval on every turn of the conversation. No lag. No dead air. We sent the video. They thought it was fake. So we deployed it on Livekit and gave them live access. Try it yourself, real-time, no tricks. They did. Then they asked us to take over their entire voice agent stack. Not just retrieval. Everything. That moment changed how I think about what we're building. We took the same approach and built a voice agent for our own website. It answers questions about Moss in real time, pulling from our docs with zero lag. Go to moss.dev and try it.

    • No alternative text description for this image
  • 𝗠𝗼𝘀𝘀 𝗶𝘀 𝗼𝗳𝗳𝗶𝗰𝗶𝗮𝗹𝗹𝘆 𝗦𝗢𝗖𝟮 𝗰𝗼𝗺𝗽𝗹𝗶𝗮𝗻𝘁 Shipping agents to enterprise usually means months of security reviews and compliance headaches. We handled the heavy lifting so you don't have to. Start building for enterprise with Moss's real-time retrieval engine now. Quick Start → https://lnkd.in/gCgC4zFY #AI #Enterprise #YC #Developers #SOC2

  • 𝗪𝗲 𝘀𝗵𝗶𝗽𝗽𝗲𝗱 𝗮 𝘃𝗼𝗶𝗰𝗲 𝗮𝗴𝗲𝗻𝘁 𝗼𝗻 𝗼𝘂𝗿 𝗵𝗼𝗺𝗲𝗽𝗮𝗴𝗲. because the best way to explain sub-10ms retrieval is to let you talk to it. Ask it anything about Moss. It pulls from our docs. You'll hear the difference. Every other retrieval layer adds latency you can feel in a conversation. Ours doesn't. moss.dev → Start Conversation → see for yourself. #VoiceAI #SemanticSearch #YCombinator #DevTools #AIInfra

  • Every developer call ended the same way. "Is it open source?" Not "show me a demo." Not "send a whitepaper." That's when it clicked. If you're building something new, developers won't just take your word for it. They need to see it, run it, break it, and trust it themselves. And for us, it went deeper than that. We're not just shipping another tool. We're pushing forward what local-first, real-time search for AI should look like. You don't build a new category in private. You build it with a community. That's where we were wrong. We stayed heads down for too long. Overthought it. Treated open source like a distraction instead of the distribution and trust layer it actually is. So we stopped overthinking and pushed the repo live. No launch. No campaign. Just shipped. What happened next: → Contributors showed up day one → Developers started building before we even posted → The community started forming on its own If this category is going to exist, it won't be because we say it should. It'll be because developers adopt it, extend it, and make it their own. Building in public isn't optional here. It's the only way forward. Curious to see what people build.

    • No alternative text description for this image
  • 𝗶𝗻𝗳𝗲𝗿𝗲𝗱𝗴𝗲-𝗺𝗼𝘀𝘀 𝗣𝘆𝘁𝗵𝗼𝗻 𝗦𝗗𝗞 𝘃𝟭.𝟬.𝟬𝗯𝟭𝟵 𝗶𝘀 𝗹𝗶𝘃𝗲 → comes with full python 3.14 support. → embedding generation is faster now. should make a noticeable difference in your pipelines. → You now get per-query routing metrics: local vs. cloud execution counts, latency breakdowns. Aggregated stats only, no query payloads leave your environment. Sneak Peek: Portal dashboards that bring all your analytics to the surface coming soon!

  • Moss (YC F25) just crossed 1,000 users. No billboards. No paid ads. No growth hacks. Just developers finding us and deciding to stay. From day one we made a bet: treat developers as first-class citizens. Not just in docs or APIs, but in how we build the entire product. Every feature starts with the question "does this make a developer's life easier?" Developers tried Moss, saw the speed, and started pulling it into their production stacks. One engineer becomes a team. One team becomes a company. That's how 1,000 happened. The best part? These aren't tire-kickers. They're teams shipping real AI products who need retrieval that just works. And they keep coming back. A thousand users in, and we're just getting started. Thank you to everyone who's built with us so far. What's next? A lot. Stay tuned.

Similar pages