Generative AI May 24, 2025

Building a Healthcare AI Prototype with NotebookLM, Gemini 2.5, Firebase, and Replit

Google’s current AI toolchain compresses a lot of early product work into a single loop: research, synthesis, code generation, lightweight deployment, even landing-page copy. The healthcare prototype in the “Blood Works” walkthrough shows how far tha...

Building a Healthcare AI Prototype with NotebookLM, Gemini 2.5, Firebase, and Replit

NotebookLM, Gemini 2.5, Firebase, Replit: a fast AI app stack that’s impressive and a little reckless

Google’s current AI toolchain compresses a lot of early product work into a single loop: research, synthesis, code generation, lightweight deployment, even landing-page copy. The healthcare prototype in the “Blood Works” walkthrough shows how far that loop now goes.

The stack is simple:

  • NotebookLM for source gathering and analysis
  • Gemini 2.5 for code generation and UI work in Canvas
  • Firebase Hosting for quick deployment
  • Replit for publishing a separate marketing site

The demo app is a centralized blood lab results dashboard with AI interpretation and sharing features. Good prototype subject. It combines a real user pain point, regulated-looking data, and a UI people can understand quickly. It also exposes how fast “built in hours” turns into “needs actual safeguards.”

What matters here

The obvious pitch is speed. Every demo says that.

What’s worth paying attention to is the workflow. NotebookLM’s newer research flow, especially the Discover-style source collection, does a lot of the work upfront. Instead of bouncing between Reddit, healthcare forums, consultant PDFs, and competitor sites, the process pulls sources into one place, labels them, and lets you query subsets of the material.

That sounds small until you’ve watched teams spend days pasting findings into docs and losing any usable structure.

The labeling scheme is basic and useful:

  • USER_1, USER_2 for user pain points
  • TRENDS_1, TRENDS_2 for industry reports
  • COMPETITOR_1, COMPETITOR_2 for market references

That gives you controlled context instead of one giant prompt blob. With a 1 million token context window, it’s easy to throw everything in and hope the model sorts it out. That usually produces mush. Toggling sources on and off and asking narrower questions is a much better pattern.

For developers and tech leads, that’s the takeaway. Source discipline matters more than the model flex.

NotebookLM as lightweight research ops

The source material presents NotebookLM as a faster alternative to using search, Perplexity, Google Scholar, and manual notes side by side. Fair enough. What stands out is that it starts behaving like a lightweight research-ops layer for early product discovery.

The prompts in the demo are practical:

Find sources on Reddit, Quora, and healthcare forums
where users discuss understanding, accessing, sharing blood lab results.

Then:

Gather studies from leading consulting firms (2024+)
on consumer healthcare app trends.

And:

List homepages of fast-growing consumer healthcare apps.

Nothing magical there. They’re solid research briefs. The advantage is lower collection friction and findings you can query later.

That matters because AI product teams often skip discovery entirely. They jump from “interesting model capability” to generated prototype, then realize a few weeks later that users don’t trust the feature, the language is off, and the UI solves a problem nobody actually described.

This workflow helps with that. It doesn’t solve it. Forum scraping and trend reports can give you an early signal, but they won’t replace interviews, domain review, or serious validation. In healthcare, that gap matters.

Gemini 2.5 changes the pace

The coding side is familiar by now, but still notable. Gemini 2.5 is used to generate a React and TypeScript prototype with a defined feature set:

  • data aggregation
  • AI-driven interpretation
  • secure sharing
  • authentication
  • basic UI

Prompts at that level now often produce a plausible file structure and code that compiles with limited cleanup. A year or two ago, “build me a working prototype” usually got you a toy. Now it can produce something coherent enough to hand to a designer, PM, or customer.

Canvas is useful here too. It gives Gemini a more visual path for UI generation, homepage layout, and infographic-style assets. Frontend teams will recognize the value immediately. Faster iteration on structure and presentation is real. So are the limits. Generated UI can look polished while still being brittle, inaccessible, or full of weird state handling.

The source notes a practical issue with icons and components failing. The suggested fix is to specify libraries like shadcn/ui and make sure Canvas is enabled. That lines up with current model behavior. If you don’t pin the component assumptions, you get generic frontend soup.

So yes, Gemini can produce a prototype quickly. The bigger shift is where the bottleneck moves next: data modeling, auth, security boundaries, and product judgment.

Firebase and Replit keep the loop short

Deployment in the walkthrough is blunt:

  1. firebase init hosting
  2. copy generated files into /public
  3. firebase deploy

For a demo, that’s enough. For a static or mostly client-side prototype, Firebase Hosting is still one of the quickest paths from generated code to a live URL. That’s useful.

Then there’s the separate move: using Replit to host the marketing homepage generated in Gemini Canvas. The prompt to Replit is basically “take this code and deploy it.”

It’s a bit messy. App on one platform, landing page on another. Still, it reflects how these AI-assisted workflows often grow. Teams take the shortest path at each step instead of stopping to define a coherent architecture. Early on, that’s fine. Later, it becomes drag.

If you’re making engineering decisions, that’s the part to watch. AI tools cut generation time. They don’t reduce integration debt. Sometimes they increase it because it becomes cheap to spin up one more half-connected artifact.

The healthcare part gets shaky fast

A bloodwork dashboard is a smart prototype subject because the value is easy to grasp. People do struggle to interpret lab results, access them across providers, and share them with family or clinicians.

It’s also where demo confidence gets ahead of reality.

AI-driven interpretation in healthcare carries obvious risk. If the prototype is just a front end with mock data, fine. If somebody starts sending real patient data into third-party model APIs, the bar changes immediately. You have to think about:

  • PHI handling and whether any data leaves your controlled boundary
  • consent and data retention rules
  • auditability of generated interpretations
  • hallucinations presented as medical guidance
  • auth and sharing controls that need to be stronger than basic login
  • regional compliance requirements, including HIPAA in the US if covered entities or business associates are involved

The source material mentions Firebase authentication errors and the need to verify firebase.json and OAuth setup. Fine as a deployment note. For anything touching health data, auth is just the first layer. Access control, secure storage, encryption, logging, and legal obligations matter far more than how quickly the landing page is live.

The prototype still has value. It just needs an honest label: a research-backed front-end prototype, not a healthcare product.

The meta-prompt pipeline is the best idea here

The walkthrough leans on meta-prompts, prompts that generate reusable prompts for later steps. That’s one of the stronger habits in this stack.

If you’re running repeated discovery and prototyping cycles, you don’t want bespoke prompting every time. You want reusable templates for:

  • source collection
  • source labeling
  • pain-point extraction
  • trend alignment
  • MVP scoping
  • code generation
  • landing-page generation

That makes the workflow repeatable across ideas and across teams. It also makes failures easier to inspect. If a prompt chain starts producing junk, you can see where it drifted.

That’s a meaningful shift. Parts of ideation, research synthesis, and prototyping can now be turned into process instead of treated like one-off model wizardry.

Where this stack fits

For hackathons, internal incubators, founder-led validation, and design engineering, this stack is genuinely strong. One person can go from market scan to live demo in an afternoon. That used to take either broad full-stack range or a small team.

For production software, the hard parts are still waiting:

  • data contracts
  • observability
  • security review
  • testing
  • maintainability
  • vendor exposure
  • regulatory risk

That’s even more true in health-adjacent software, where a slick prototype can hide a lot of unresolved danger.

The demo does get one thing right. It treats AI tooling as a pipeline. NotebookLM narrows the problem. Gemini drafts the product. Firebase and Replit publish the result. For MVP work, that’s a practical stack. For software that has to survive compliance review, it’s an early draft and nothing more.

These tools can compress the first 20 percent of product work aggressively enough that teams forget the other 80 percent is still there.

Keep going from here

Useful next reads and implementation paths

If this topic connects to a real workflow, these links give you the service path, a proof point, and related articles worth reading next.

Relevant service
AI model evaluation and implementation

Compare models against real workflow needs before wiring them into production systems.

Related proof
Internal docs RAG assistant

How model-backed retrieval reduced internal document search time by 62%.

Related article
Google NotebookLM adds Video Overviews to turn specs and docs into slide decks

Google is rolling out Video Overviews for NotebookLM. You drop in a spec, RFC, white paper, meeting notes, or a pile of markdown files, and NotebookLM generates a narrated slide deck from those sources. Google says it starts rolling out on desktop, i...

Related article
OpenAI's ChatGPT Memory with Search rewrites web queries from chat history

OpenAI is moving ChatGPT further from the old search-box model. The new feature, Memory with Search, lets ChatGPT use details saved from past conversations to rewrite web queries before they go out. Ask for “restaurants near me,” and if it knows you’...

Related article
Gemini 2.5 Pro vs Flash: What Google I/O 2025 changed for developers

Google used I/O 2025 to push a lot of AI product demos. The part that matters more is the developer story around Gemini 2.5. For once, it feels clearer. The model lineup is simple enough: Gemini 2.5 Pro is the top-end model, Gemini 2.5 Flash is the c...