Skip to content
View JosephGabito's full-sized avatar
🔥
Grilling Chicken Inasal
🔥
Grilling Chicken Inasal

Organizations

@codeablehq @UncannyOwl @dunhakdissoftwarecreatives

Block or report JosephGabito

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
JosephGabito/README.md

About

I architect production AI systems.

For over 16 years, I’ve built scalable backend infrastructure. Today, I design deterministic systems around probabilistic models: multi-agent orchestration platforms, retrieval infrastructure, and distributed execution engines that operate reliably under load.

I don’t build AI demos. I build systems that survive production.

My work sits at the boundary between stochastic language models and deterministic software architecture. That boundary is where most systems fail and where real engineering matters.

Core domains:

  • Multi-agent orchestration and execution engines
  • Retrieval-Augmented Generation (RAG) infrastructure
  • Distributed task systems and background processing
  • AI evaluation harnesses and reliability tooling
  • Schema-first backend architecture using FastAPI, Celery, Redis, and PostgreSQL

Generative models are probabilistic. Infrastructure must not be.

Architecture Focus

🧠 Agent Systems

  • Multi-agent coordination graphs and execution loops
  • Tool routing and structured output enforcement
  • Streaming pipelines with background task isolation
  • Failure recovery, retry policies, and state reconciliation
  • Long-running workflow orchestration

Designed to eliminate silent failure modes and nondeterministic behavior in AI-driven systems.

📚 Retrieval Infrastructure

  • Hybrid search pipelines combining sparse and dense retrieval
  • Embedding normalization and scoring strategies
  • Retrieval evaluation using precision, recall, and nDCG
  • FAISS to cuVS experimentation and performance benchmarking
  • Hallucination risk mitigation through retrieval grounding

Built to transform retrieval quality from intuition into measurable signal.

⚙️ Distributed Backend Systems

  • Asynchronous task orchestration and worker pools
  • Message brokers and job queues
  • Schema-first API contracts and strict boundaries
  • Observability, tracing, and load diagnostics
  • Deterministic control planes around AI components

Focused on reliability, not novelty.

Selected Work

Production Agent Orchestration Platform

Architected and deployed a multi-agent execution system integrating:

  • Tool routing
  • RAG context injection
  • Structured outputs
  • Streaming and background processing
  • Failure handling and idempotent retries

Built with FastAPI, Celery, Redis, and PostgreSQL. Designed for concurrency, resilience, and long-running execution flows.

RAG Evaluation and Reliability Harness

Engineered automated pipelines to:

  • Measure retrieval quality
  • Compare embedding and scoring strategies
  • Detect hallucination risk patterns
  • Benchmark latency and system stability under stress

Focused on bridging research metrics with production guarantees.

Certifications

  • NVIDIA AI Certified: Generative AI
  • NVIDIA AI Certified: Agentic AI Applications with Large Language Models
  • Codeable Certified WordPress Expert

Engineering Philosophy

  • Explicit over magical
  • Boring systems are better than clever hacks
  • Deterministic boundaries around probabilistic models
  • Clean architecture enables safe iteration
  • Production reliability is the benchmark

Connect

LinkedIn: https://www.linkedin.com/in/joseph-gabito/

Email: dsc [dot] official [dot] mail at gmail [dot] com

Pinned Loading

  1. Single-file training script for a GP... Single-file training script for a GPT-style decoder trained on TinyStories using GPT-2 BPE (tiktoken). Implements a modern pre-norm stack (RMSNorm + causal SDPA + GELU MLP) with weight tying, bf16 mixed precision (when available), warmup + cosine LR decay, gradient clipping, checkpointing (best + periodic), and a built-in sample generation at the end.
    1
    """
    2
    i16 — Train a GPT-style decoder on TinyStories (tiktoken GPT-2 BPE).
    3
    4
    Usage:
    5
      python i16_train.py --data TinyStoriesV2-GPT4-train.txt --out checkpoints