Back to Projects

GovChime Analytics Platform

Government Contracts Intelligence

Overview

7-package monorepo powering a government contracts intelligence platform. Architected 3 SmartSync microservices (Data, Refresh, ClickHouse sync) for multi-source data synchronization across SamgovAPI, PostgreSQL, and ClickHouse OLAP. Next.js frontend deployed on Cloudflare Workers, Express.js API on Komodo, with 24+ GitHub Actions CI/CD workflows (unit → integration → E2E → deploy) running on a self-hosted runner. AI pipeline handles data sanitization, contract matching, and description generation across 70M+ row datasets.

Key Features

  • 7-package monorepo with 3 SmartSync microservices (Data, Refresh, ClickHouse sync)
  • 50+ materialized views reducing query times by 100-300% on 70M+ row datasets
  • Next.js frontend on Cloudflare Workers with ISR, SLED Admin dashboard on Cloudflare Pages
  • 24+ GitHub Actions CI/CD workflows on self-hosted runner with Komodo deployment orchestration
  • AI pipeline for automated data sanitization, contract matching, and description generation
  • Stripe paywall integration, PostHog analytics, and TDD-driven agentic AI development workflow

Tech Stack

Frontend

Next.jsReactTypeScriptTailwind CSSCloudflare Workers

Backend

Express.jsNode.jsTypeScriptREST APIStripe

Database & OLAP

PostgreSQLClickHouseMaterialized ViewsOLAP

Infrastructure

DockerKomodoGitHub ActionsCloudflare PagesSelf-Hosted Runner

AI & Dev Tools

OpenAI APIClaude CodeMCPPlaywright

Challenges & Solutions

Slow Analytics Queries on 70M+ Rows

Problem

Real-time aggregation queries across 70M+ rows with complex JOINs took seconds, making dashboards unusable for end users.

Solution

Designed ClickHouse OLAP integration alongside PostgreSQL with 50+ materialized views for common aggregations. Query times reduced by 100-300% — dashboards became instant.

Multi-Service CI/CD for Sole Engineer

Problem

7 packages with interdependent builds and deploys needed reliable CI/CD without a dedicated DevOps team. Frontend ISR depends on backend being live, services must deploy atomically.

Solution

Architected 24+ GitHub Actions workflows on a self-hosted runner with dynamic port allocation, Komodo HTTP API for Docker orchestration, and a Build → Verify → Deploy pipeline ensuring frontend validates against temp backend before any production deploy.

Data Quality at Scale

Problem

Raw SamgovAPI data contained inconsistencies, missing fields and unstructured descriptions making it difficult to search and match contracts.

Solution

Built AI pipeline using LLM APIs for automated data sanitization, opportunity matching and description generation. Structured output validation ensures consistent data quality.

Key Achievements

100-300%
Query time reduction via ClickHouse OLAP + materialized views
70M+ Rows
Production dataset with real-time analytics
7 Packages
Monorepo with 3 SmartSync microservices + API + frontends
24+ Workflows
CI/CD pipeline: unit → integration → E2E → deploy