Skip to main content

23 posts tagged with "ai"

View All Tags

Introducing Parthenon: Transforming Healthcare with AI-Powered Outcomes Research

· 6 min read
Creator, Parthenon

Pinned Post | Originally published March 7, 2026

Outcomes research has evolved alongside the broader arc of healthcare analytics infrastructure. Early siloed clinical systems produced fragmented administrative and claims data with limited analytic utility — adequate for billing, but structurally unsuitable for longitudinal cohort construction or comparative effectiveness work. The meaningful use era expanded the availability of structured clinical data, yet interoperability failures meant that patient journeys remained fractured across institutional boundaries, undermining the real-world evidence studies that outcomes researchers depend on. The shift to integrated analytics platforms — particularly the adoption of common data models like OMOP/OHDSI — marked a genuine inflection point: federated network studies, standardized phenotyping, and reproducible retrospective analyses became operationally feasible at scale. Now a fourth generation is taking shape, one in which AI-augmented clinical intelligence moves outcomes research from retrospective description toward prospective, near-real-time evidence generation — enabling dynamic cohort surveillance, treatment heterogeneity detection, and value-based care signal identification that was previously impractical outside of narrow clinical trial settings.

Parthenon is built for this fourth generation.

Jobs Page Overhaul, Drug Era Performance Breakthrough, and Cohort Pipeline Hardening

· 5 min read
Creator, Parthenon
AI Development Assistant

A landmark day for platform observability and data pipeline reliability. We shipped a fully wired Jobs monitoring page that surfaces all 13+ tracked job types, broke through a major ETL performance ceiling on the SynPUF dataset (17 hours → 14 minutes for drug_era builds), and closed out a cohort generation audit that uncovered eight discrete bugs across the SQL builders, API layer, and frontend.

One Million Patient Embeddings: GPU-Accelerated Similarity Search Comes to Parthenon

· 20 min read
Creator, Parthenon
AI Development Assistant

Two days ago, we shipped the Patient Similarity Engine — a multi-modal system that scores patients across six clinical dimensions on OMOP CDM. The architecture was sound. The algorithms worked. But there was a problem hiding in plain sight: none of our patients had embeddings.

The embedding pipeline had been silently failing since day one. Three type mismatches between our PHP backend and Python AI service meant that every embedding request returned a validation error, was caught by a try/catch block, and logged as a warning that nobody read. The feature vectors were all there — conditions, drugs, measurements, procedures — but the 512-dimensional dense vectors that would make similarity search fast at scale? Zero. For every source. For every patient.

Tonight, we fixed all three bugs, refactored the embedding pipeline from CPU-only SapBERT to GPU-accelerated Ollama, upgraded from 512 to 768 dimensions, introduced batch deduplication that delivered a 123x throughput improvement, and generated embeddings for 1,007,007 patients across three CDM sources. This is the story of what broke, what we built, and what it unlocks.

Patients Like Mine: Building a Multi-Modal Patient Similarity Engine on OMOP CDM

· 18 min read
Creator, Parthenon
AI Development Assistant

For twenty years, the question "which patients are most like this one?" has haunted clinical informatics. Molecular tumor boards want to know: of the 300 patients in our pancreatic cancer corpus, which ones had the same pathogenic variants, the same comorbidity profile, the same treatment history — and what happened to them? Population health researchers want to seed cohort definitions not from abstract inclusion criteria but from a concrete index patient. And every clinician who has ever stared at a complex case has wished for a button that says show me others like this.

Today, Parthenon ships that button. The Patient Similarity Engine is a multi-modal matching system that scores patients across six clinical dimensions — demographics, conditions, measurements, drugs, procedures, and genomic variants — with user-adjustable weights, dual algorithmic modes, bidirectional cohort integration, and tiered privacy controls. It works across any OMOP CDM source in the platform, from the 361-patient Pancreatic Cancer Corpus to the million-patient Acumenus CDM.

This post tells the story of why it was needed, what we studied before building it, how it works under the hood, and what we learned along the way.

The Magical Ladies of Parthenon

· 11 min read
Creator, Parthenon
AI Development Assistant

In Greek mythology, the great temple atop the Acropolis housed not just Athena, but an entire pantheon of divine figures — each wielding a unique gift. Parthenon, our unified OHDSI outcomes research platform, follows the same philosophy. Behind the scenes, four mythological women power the intelligence layer that transforms raw clinical data into actionable research: Hecate, Phoebe, Ariadne, and Arachne.

CI Green at Last: Codebase Hardening, AtlanticHealth Synthesis, and a 147-Test Renaissance

· 5 min read
Creator, Parthenon
AI Development Assistant

After months of a perpetually red CI pipeline, today marks a turning point for Parthenon: 92 commits, a full-spectrum codebase review, a complete AtlanticHealth patient synthesis pipeline, and — most satisfying of all — every CI job green. Here's how we got there.

Abby 2.0 Phase 5: Advanced Agency — Parallel Workflows and Safety Rails

· 3 min read
Creator, Parthenon
AI Development Assistant

Abby can now orchestrate complex multi-step research workflows with independent steps running in parallel. High-risk tools (modify concept sets, update cohort criteria, execute SQL) join the toolkit with safety validation. Dry run mode simulates actions before execution. Workflow templates encode OHDSI best practices into one-click study designs.

Abby 2.0 Phase 4: The Agency Framework — She Gets Hands

· 5 min read
Creator, Parthenon
AI Development Assistant

Abby can now take actions, not just answer questions. "Build me a diabetes cohort" generates a reviewable multi-step plan — create concept sets, define the cohort, generate the patient count — that executes with one click after user approval. Every action is logged with checkpoint data for rollback. Phase 4 adds supervised autonomy with safety rails.

Abby 2.0: From Chatbot to Cognitive Research Assistant — The Complete Architecture

· 15 min read
Creator, Parthenon
AI Development Assistant

In a single development session, we shipped three phases of a cognitive architecture that transforms Abby from a stateless RAG chatbot into a persistent, intelligent, context-aware research assistant. She now remembers who you are, routes complex questions to a more powerful brain, traverses clinical concept hierarchies, and warns you when your data has gaps. This post tells the complete story — the problems we solved, the architecture we built, and the engineering decisions behind 188 passing tests across 60+ new files.

Abby 2.0 Phase 6: Institutional Intelligence — The Organization Gets Smarter

· 4 min read
Creator, Parthenon
AI Development Assistant

Abby now learns from the entire research community. When a researcher builds a successful diabetes cohort, that pattern becomes available to every other researcher. Questions asked three or more times across users automatically become institutional FAQs with vetted answers. Data quality findings discovered by one team warn all teams. Phase 6 completes the Abby 2.0 cognitive architecture.