CASE STUDIES

SaaS, AI, health care, and financial startups trust FivexL to build their infrastructure in AWS, empowering their businesses to grow faster. Learn how. Book a consultation
 

Lead consultant for the project

Ivan.png

Ivan Razzhivin

Certified AWS Architect AWS ECS and Terraform Expert
savehealth_logo.png

About SaveHealth

SaveHealth is a U.S.-based healthcare startup. Their product helps patients and consumers find lower prescription drug prices by comparing costs across pharmacies and accessing discount coupons and cards. It is especially useful for people paying out of pocket, without insurance, or looking to compare discount prices with their insurance copay. SaveHealth solves the problem of high and inconsistent prescription drug prices across the U.S. market.

Challenge

SaveHealth had a clear strategic goal: expose proprietary internal data to leading external LLM ecosystems - primarily OpenAI and Anthropic - in a secure, production-grade way.

This was not a "build a chatbot" engagement. The real challenge was to create an AI integration layer that could securely expose sensitive healthcare data through an MCP-compatible interface, work with fast-moving external AI providers and their evolving integration requirements, remain scalable and maintainable in production, and give the team a reusable platform for future AI initiatives - not just a one-off proof of concept.

The business objective was to unlock a new traffic and product-distribution channel through LLM platforms. The technical challenge was doing this without compromising security, operational simplicity, or AWS architectural standards - all while remaining HIPAA-compliant.

A major complexity was that the AI integration landscape was evolving during the project itself. Requirements from providers such as OpenAI were still maturing, which meant the solution had to be both production-ready and adaptable.

SaveHealth already had a solid conventional AWS setup in place - a frontend layer, CloudFront for caching and request acceleration, backend APIs running on ECS, and managed AWS data-processing components such as Airflow and Glue jobs. The foundation was there. What was missing was an AI-native platform layer that could securely connect internal systems and proprietary data to external LLM providers in a structured, scalable, and standards-aligned way.

Solution

FivexL designed a staged approach: validate first with a pragmatic PoC, then industrialize into a production-grade AI platform - adapting the architecture as the ecosystem matured in real time.

Rapid PoC on ECS
Rather than waiting for the AI ecosystem to stabilize, FivexL started with a pragmatic proof of concept on ECS to validate the MCP integration model quickly. This let SaveHealth test both the business and technical viability of exposing internal data to external LLM platforms immediately. The backend was implemented using Python and FastMCP, optimizing for speed of execution, protocol alignment, and developer productivity.

Production-Grade CI/CD from Day One
FivexL built a production-style CI/CD pipeline with GitHub Actions from the beginning, so the project would not become a fragile prototype. The pipeline covered build, delivery, and validation workflows. Docker image build and delivery into AWS ECR was automated with ECR Scanning for CVEs, ensuring a clean, repeatable artifact flow. End-to-end testing was added early to reduce regression risk across the integration lifecycle.

Infrastructure as Code
All infrastructure was defined using Terraform, so cloud resources, environments, and deployment paths could be reproduced consistently and safely - following AWS best practices around multi-account architecture, security guardrails, and platform standardization.
Evolution to Managed AI Platform
Once AWS released AgentCore on Bedrock, FivexL reassessed the architecture and intentionally evolved the solution from a self-managed to a managed MCP hosting model. AgentCore materially reduced operational burden, improved long-term maintainability, and better matched SaveHealth’s strategic need: not just to run one server, but to establish a scalable AI platform capability.

Modern Authentication with OAuth 2.1
FivexL leveraged OAuth 2.1 support through AgentCore as part of the managed architecture, aligning the solution with the emerging standard for MCP server integrations and reducing security implementation friction.

Secure Data Exposure
The platform was designed to securely expose proprietary internal data while keeping sensitive systems and control planes inside AWS - addressing both SaveHealth’s security posture and the practical realities of healthcare AI integration.

The final outcome was not just an integration path, but a scalable AI platform layer that lets development teams move faster, launch experiments more safely, and bring AI workloads to production without rebuilding infrastructure every time.

Benefits

From concept to production-grade AI platform - SaveHealth went from "we need to connect our internal data to OpenAI and Anthropic securely" to having an AWS-native AI platform that supports production-grade MCP integrations, future AI workloads, and faster experimentation with lower operational overhead.

Immediate outcomes (first 2-4 weeks)
The MCP-based architecture was validated through a working PoC. SaveHealth gained a functioning backend integration layer for testing against external LLM providers. CI/CD, container delivery, testing, and Terraform-based infrastructure workflows were established early. The project moved from abstract AI strategy into executable platform capability quickly.

Longer-term outcomes
The architecture matured from a PoC runtime into a reusable AI platform foundation. Operational complexity was reduced through migration toward a managed AgentCore-based model. SaveHealth gained a cleaner and more cost-effective long-term operating model than an ECS-heavy approach. Internal teams can now launch future AI workloads faster using the same platform base.

What the client got beyond expectations
Not just an MCP server or an isolated AI integration - but a real AI platform foundation: a reusable AWS-native pattern for future AI workloads, a secure method for exposing internal data to external LLM ecosystems, a lower-maintenance operating model, and a production-ready environment where teams can move from experiment to deployment much faster. It converts AI work from bespoke engineering effort into platform capability.
Start Your Success Story

Our happy customers