Skip to content
View BlaiseMoses01's full-sized avatar

Block or report BlaiseMoses01

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
BlaiseMoses01/README.md

Blaise Moses

AI Engineer • Backend Systems • Developer Tooling

Building reliable AI infrastructure, evaluation systems, and high-leverage internal tooling.

PortfolioLinkedInEmailYouTube


About Me

I’m a Software Engineer focused on building observable, trustworthy, and scalable AI systems.

I currently work at a Fortune 500 financial institution, where I build tooling at the intersection of LLM systems, retrieval pipelines, data validation, and engineering infrastructure. My background in large-scale automation and reliability engineering has given me a strong skillset in measurement, correctness, and production confidence, as well as a deep understanding of distributed & applied AI systems.

I’m especially interested in:

  • RAG systems and evaluation pipelines
  • AI infrastructure, observability, and governance
  • Developer tooling and workflow automation
  • Agentic systems and MCP-style orchestration
  • High-leverage backend engineering problems

Anything worth doing is worth doing all the way.
Don’t mistake a plateau for the peak.


Currently Building

  • RAG evaluation and validation tooling for model-driven systems
  • Internal analytics and monitoring workflows for AI quality and consistency
  • Agentic QA and automation experiments using Playwright + Claude Code
  • MCP-style integrations for safer, more capable AI workflows
  • Personal engineering content around Claude Code, tooling, and systems thinking

Experience

AI Engineer

November 2025 – Present

  • Design and build evaluation tooling for retrieval-augmented and model-driven systems, with an emphasis on making AI behavior measurable, explainable, and production-ready.
  • Develop RAG evaluation pipelines using metrics such as Precision@k, Recall@k, and MRR to better understand retrieval quality, ranking performance, and answer usefulness.
  • Build internal analytics and validation tools with Python, Pandas, and Streamlit to surface trends in model quality, retrieval effectiveness, and system consistency.
  • Contribute to drift detection and monitoring workflows that identify changes in retrieval behavior, response quality, and data integrity across enterprise AI systems.
  • Create automated validation frameworks that cross-reference outputs across S3 artifacts, Salesforce data, internal services, and structured evaluation datasets to improve traceability and confidence.
  • Explore agentic engineering patterns and MCP-enabled workflows for safely connecting models to internal tools, development pipelines, and evaluation systems.

Software Engineer (Test Infrastructure / Reliability)

October 2024 – November 2025

  • Built and scaled UI and API automation infrastructure using Playwright, TypeScript, and Cucumber for enterprise applications with high reliability requirements.
  • Designed supporting orchestration layers and backend workflows that made automation more deterministic, maintainable, and scalable.
  • Developed CI/CD regression systems with GitLab CI, AWS Lambda, DynamoDB, and S3, improving pipeline reliability and reducing operational friction.
  • Served as a technical owner for framework architecture, reliability improvements, and engineering standards around automation and validation.
  • Conducted technical interviews and helped onboard engineers into shared tooling, framework usage, and quality-focused engineering practices.

Focus Areas

AI Systems
RAG Evaluation Precision@k Recall@k MRR Drift Detection Model Validation

Backend & Infrastructure
AWS Event-Driven Systems Data Pipelines System Reliability Internal Tooling

Developer Tooling
Playwright FastAPI FastMCP Claude Code Automation Frameworks

Languages
Python TypeScript JavaScript C++ Java SQL


Featured Projects

Agentic Playwright Team

A multi-agent browser automation experiment focused on systematic exploration, test planning, and accelerated Playwright authoring.

Repo

Claude Code from Zero

A practical YouTube course covering CLI basics, context, CLAUDE.md, skills, sub-agents, and MCP.

PlaylistChannel

Melton’s Home & Auto

A production Next.js web app and digital presence project for my family’s automotive business.

Live Site

Portfolio

A central hub for my projects, writing, and technical explorations across AI, systems, and software engineering.

Website


Connect

Popular repositories Loading

  1. agentic-playwright-team agentic-playwright-team Public

    Agentic Testing Automation Workflow using Claude Agent Teams, Vercel Agent-Browser CLI, and Playwright/CucumberBDD

    TypeScript 1

  2. two-way-asynchronous-messenger two-way-asynchronous-messenger Public

    a two way asynchronous messenger console app written in Python

    Python

  3. groupchat_messenger groupchat_messenger Public

    a group chat server/client console application

    Python

  4. cracker_barrel_pin_game cracker_barrel_pin_game Public

    exhaustive search algorithm I designed to find all possible solutions to the Cracker Barrel Pin game from every starting position

    Python

  5. FCC_ML_CertProjects FCC_ML_CertProjects Public

    This repository contains the code for all of the projects I completed to earn my Machine Learning in Python Certification from FreeCodeCamp.org

    Jupyter Notebook

  6. NeatFiles NeatFiles Public

    a tkinter python app used for sorting file directories by extension

    Python