Participate in our State of AI Development Surveyfor a chance to win a MacBook M4 Pro!Take 4-min Survey →
Case Studies
Miri: Collaboratively building a chatbot in production with Vellum
Oct 13, 2023
Akash Sharma
Co-founder & CEO
Co-authors:
No items found.
Table of Contents

Recent advances in Large Language Models have fueled incredible new chat experiences where end users can converse with AI. However, building a custom, production-grade, AI-powered chat experience is still fraught with challenges.

In this post, we break down how Miri, a company specializing in AI-powered wellness coaching, approached the problem and the successful solution they ended with (try out Miri and get your personal wellness companion for free here if you’re curious!)

Who is Miri, and what brought them to Vellum?

Miri is an early-stage startup aiming to revolutionize the health and wellness industry. Miri allows people from around the world to get access to the top health experts at a much lower cost through a personalized ongoing coaching experience. Chat is a cornerstone of this AI coaching product and is built using LLMs to mimic the experience individuals get when they work directly with top experts; a deep understanding of their personal history, education on key topics, personalized recommendations for behavior change, and thoughtful check-ins.   

When we first met the team at Miri, their first priority was to deeply integrate LLMs into their product. Before coming to Vellum, the team had built an internal proof of concept which was really powerful and showed the potential of LLMs for this use case but did not give them the flexibility they needed for production use. Popular open-source libraries weren’t cutting it for their use cases given their bespoke needs. Product Advisor Adam Daigian knew they wanted significant flexibility to tweak instructions given to the model (for better quality) and test different product experiences without having to iterate exclusively at the codebase level and through Colab/Jupyter notebooks. Time to market was of the essence and the cross-functional team needed a platform to collaboratively build and iterate on their LLM powered chat experience.

How Miri uses Vellum for their product development?

When the team at Miri was looking for a platform to build, they wanted a place to rapidly test, iterate, and deploy multiple prompts. They were looking to work closely with a team that understood their requirements and were willing to invest in product capabilities to support their development. The Vellum team provided architecture advice and prompt engineering best practices to quickly onboard the product and engineering team at Miri.

The LLM development process at Miri has been transformed in the last few months. Using LLMs in production is no longer a blackbox and new ideas/use cases can go into production within a week (with sufficient testing!). Engineering time is effectively leveraged and used primarily for final testing, integrations, and building a magical end user experience. While the Miri team started by trying to solve the initial LLM development problem, what they’ve really enjoyed about using Vellum is that it’s proving to scale with the team’s needs once the chat experience is in production too.

Since the LLM chatbot and underlying logic (how prompts, conditional logic, semantic search all tier together) are all built and deployed in Vellum, the team at Miri is able to quickly spot and diagnose edge cases in production. For a recent edge case they discovered in production, Adam was able to find the completion, look at the inputs & outputs at each step, isolate the problem to a specific instruction in one prompt (which was part of a multi-step prompt chain), update the instruction and redeploy the chatbot with new instructions all in under 2 hours and without having to write a single line of code. The chatbot was then able to handle this edge case in production.

Miri is rapidly iterating on its production chat experience, identifying test cases based on user responses, and continuously improving the quality of their product thanks in part to Vellum’s platform. Actual data in production can easily be added as a test case for further iteration.

What impact has this partnership had on Miri?

Vellum’s platform is an essential tool which helps Miri deliver a cutting edge consumer facing generative AI product. The fact that testing, iteration, and monitoring can happen in one place and everyone can collaborate in the development process is of enormous value to them. The collaboration is both asynchronous (independently working on the same prompt) and also synchronous (Vellum UI is shared on a screen during a call). The team at Miri is excited about multiplayer mode (a-la Google Docs & Figma) for live prompt editing to come to Vellum’s product soon!

These days, Adam, as a product advisor, is able to build 60-70% of any new conversation flow himself; The team at Miri frequently shares feature requests which would improve their development processes and have seen Vellum’s product evolve to better suit their needs.

Want to try for yourself?

Miri has a public beta you can use for free today. Beta access available now for free ongoing wellness coaching, personalized for your unique health goals.

Dozens of companies like Miri have improved the quality of their LLM applications and internal development processes when working with Vellum. Vellum provides the tooling layer to experiment with prompts and models, evaluate their quality, and make changes with confidence once in production — no custom code needed! Request a demo for our app here, join our Discord or reach out to us at support@vellum.ai if you have any questions. We’re excited to see what you and your team builds with Vellum next!

ABOUT THE AUTHOR
Akash Sharma
Co-founder & CEO

Akash Sharma, CEO and co-founder at Vellum (YC W23) is enabling developers to easily start, develop and evaluate LLM powered apps. By talking to over 1,500 people at varying maturities of using LLMs in production, he has acquired a very unique understanding of the landscape, and is actively distilling his learnings with the broader LLM community. Before starting Vellum, Akash completed his undergrad at the University of California, Berkeley, then spent 5 years at McKinsey's Silicon Valley Office.

No items found.
The Best AI Tips — Direct To Your Inbox

Latest AI news, tips, and techniques

Specific tips for Your AI use cases

No spam

Oops! Something went wrong while submitting the form.

Each issue is packed with valuable resources, tools, and insights that help us stay ahead in AI development. We've discovered strategies and frameworks that boosted our efficiency by 30%, making it a must-read for anyone in the field.

Marina Trajkovska
Head of Engineering

This is just a great newsletter. The content is so helpful, even when I’m busy I read them.

Jeremy Hicks
Solutions Architect
Related Posts
View More

Experiment, Evaluate, Deploy, Repeat.

AI development doesn’t end once you've defined your system. Learn how Vellum helps you manage the entire AI development lifecycle.