I open sourced a production MLOps pipeline. Here is what it took to get it to PyPI and Hugging Face in one day.
<p>I have been running ML pipelines in production for few years. Tens of millions of predictions a day, real money on the line, no tolerance for guesswork.</p> <p>PulseFlow started as something I built for myself. A reference architecture I kept recreating from scratch at every company because nothing open source matched what production actually demands.</p> <p>Today I packaged it, published it to PyPI, and put a live demo on Hugging Face. Here is what it covers and how to run it in under ten minutes.</p> <h2> What PulseFlow is </h2> <p>A production-grade MLOps pipeline you can clone and run immediately. Not a tutorial. Not a toy dataset. A real stack.<br> </p> <div class="highlight js-code-highlight"> <pre class="highlight shell"><code>pip <span class="nb">install </span>pulseflow-mlops <
I have been running ML pipelines in production for few years. Tens of millions of predictions a day, real money on the line, no tolerance for guesswork.
PulseFlow started as something I built for myself. A reference architecture I kept recreating from scratch at every company because nothing open source matched what production actually demands.
Today I packaged it, published it to PyPI, and put a live demo on Hugging Face. Here is what it covers and how to run it in under ten minutes.
What PulseFlow is
A production-grade MLOps pipeline you can clone and run immediately. Not a tutorial. Not a toy dataset. A real stack.
pip install pulseflow-mlops
Enter fullscreen mode
Exit fullscreen mode
Five components wired together:
-
ETL pipeline: ingestion and preprocessing with Pandas and SQLAlchemy
-
Training pipeline: model training with MLflow experiment tracking
-
Deployment service: FastAPI microservice for real-time inference
-
Orchestration: Apache Airflow DAGs for end-to-end automation
-
Full Docker Compose stack: one command to run everything
The architecture
Every enterprise ML system I have built follows the same pattern. Raw data in, predictions out, everything in between observable and reproducible.
Raw Data → ETL → Feature Store → Training → MLflow Registry → FastAPI → Clients ↑ Airflow SchedulerRaw Data → ETL → Feature Store → Training → MLflow Registry → FastAPI → Clients ↑ Airflow SchedulerEnter fullscreen mode
Exit fullscreen mode
PulseFlow makes this concrete with actual code, not diagrams.
Run it locally in four commands
git clone https://github.com/anilatambharii/PulseFlow.git cd PulseFlow python -m venv .venv && source .venv/bin/activate pip install -r requirements.txtgit clone https://github.com/anilatambharii/PulseFlow.git cd PulseFlow python -m venv .venv && source .venv/bin/activate pip install -r requirements.txtEnter fullscreen mode
Exit fullscreen mode
Then run each stage:
python etl/data_ingestion.py python etl/data_preprocessing.py python training/train_model.py uvicorn deployment.app.main:app --reloadpython etl/data_ingestion.py python etl/data_preprocessing.py python training/train_model.py uvicorn deployment.app.main:app --reloadEnter fullscreen mode
Exit fullscreen mode
MLflow logs to ./mlruns locally. No server required. If you want the full UI:
mlflow ui --port 5000
Enter fullscreen mode
Exit fullscreen mode
Or bring up the complete stack:
docker-compose up --build
Enter fullscreen mode
Exit fullscreen mode
Why I built this as open source
Three reasons.
First, I kept seeing junior engineers spend weeks building pipeline scaffolding that should take days. PulseFlow collapses that to a git clone.
Second, enterprise ML has a credibility problem with open source. Most OSS ML projects are notebooks or toy pipelines. PulseFlow is the kind of code I would put in front of a Duke Energy production environment.
Third, I am building ARGUS-AI alongside this. ARGUS is an LLM observability platform that evaluates every model output across six dimensions: Groundedness, Accuracy, Reliability, Variance, Inference Cost, Safety. PulseFlow is what you run your models through. ARGUS is how you know they are not degrading in production.
They compose. PulseFlow trains and serves. ARGUS monitors and evaluates.
What is in the repo
PulseFlow/ ├── etl/ # Data ingestion and preprocessing ├── training/ # Model training with MLflow tracking ├── deployment/ # FastAPI inference service ├── airflow/ # Orchestration DAGs ├── models/ # Model artifacts ├── ci_cd/ # GitHub Actions workflows ├── docker-compose.yml # Full stack in one command └── pyproject.toml # pip install pulseflow-mlopsPulseFlow/ ├── etl/ # Data ingestion and preprocessing ├── training/ # Model training with MLflow tracking ├── deployment/ # FastAPI inference service ├── airflow/ # Orchestration DAGs ├── models/ # Model artifacts ├── ci_cd/ # GitHub Actions workflows ├── docker-compose.yml # Full stack in one command └── pyproject.toml # pip install pulseflow-mlopsEnter fullscreen mode
Exit fullscreen mode
Live demo on Hugging Face
You can run the full ETL, training, and inference pipeline without installing anything:
PulseFlow MLOps Demo on Hugging Face Spaces
Three tabs. Load sample data, configure hyperparameters, run inference against the FastAPI endpoint simulation. All in the browser.
The production gap no one talks about
Most MLOps content stops at "train a model and log it to MLflow." That is maybe 20 percent of what production demands.
The other 80 percent:
-
What happens when your data source schema changes at 2 AM?
-
How do you roll back a model that passed validation but is failing on live traffic?
-
Who gets paged when inference latency exceeds SLA?
-
How do you prove to your compliance team that the model version in production matches what was approved?
PulseFlow gives you the structural patterns to answer all of these. It does not answer them for you because every organization's answers are different. But it gives you the right skeleton.
What I am adding next
-
LangChain integration for LLM pipeline orchestration
-
ARGUS-AI integration for automatic G-ARVIS scoring on inference outputs
-
Kubernetes deployment manifests (production-grade, not tutorials)
-
Prometheus metrics endpoint on the FastAPI service
Connect
-
GitHub: github.com/anilatambharii/PulseFlow
-
PyPI: pypi.org/project/pulseflow-mlops
-
ARGUS-AI (the observability layer): github.com/anilatambharii/argus-ai
-
Hugging Face: huggingface.co/AmbhariiLabs
-
LinkedIn newsletter Field Notes: Production AI: linkedin.com/in/anilsprasad
If you are building ML systems in production and running into the gaps PulseFlow addresses, reach out. This is open source because I want it to be the reference architecture the community builds on.
28 years of production AI. All opinions are mine. All lessons were expensive.
HumanWritten #ExpertiseFromField
---
Step 4 — Publish settings
- Series: Leave blank for now (or create "Ambharii Labs Open Source" series later)
- Schedule: Publish immediately — Tuesday 9 AM ET is ideal but today is fine given the momentum
- Click Publish
Step 5 — After publishing, copy the URL and do these immediately`
Enter fullscreen mode
Exit fullscreen mode
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modeltrainingversion
My first impression after testing Gemma 4 against Qwen 3.5
​ I have been doing some early comparisons between Gemma 4 and Qwen 3.5, including a frontend generation task and a broader look at the benchmark picture. My overall impression is that Gemma 4 is good. It feels clearly improved and the frontend results were actually solid. The model can produce attractive layouts, follow the structure of the prompt well, and deliver usable output. So this is definitely not a case of Gemma being bad. That said, I still came away feeling that Qwen 3.5 was better in these preliminary tests. In the frontend task, both models did well, but Qwen seemed to have a more consistent edge in overall quality, especially in polish, coherence, and execution of the design requirements. The prompt was not trivial. It asked for a landing page in English for an advanc
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Products

Sanctuary AI s robotic hand demonstrates zero-shot in-hand manipulation
Sanctuary AI said its robotic hand and AI system achieved the target orientation 10 times in a row without dropping the cube. The post Sanctuary AI s robotic hand demonstrates zero-shot in-hand manipulation appeared first on The Robot Report .

Report: Apple Testing AI-Powered ‘Alternative Words’ Feature for iPhone Keyboard
Apple is reportedly testing a smarter iPhone keyboard for iOS 27, with AI-powered word suggestions and improved autocorrect to enhance typing. The post Report: Apple Testing AI-Powered ‘Alternative Words’ Feature for iPhone Keyboard appeared first on TechRepublic .



Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!