Thanks to visit codestin.com
Credit goes to github.com

Skip to content
View othnielObasi's full-sized avatar
  • UnitedKingdom
  • 00:03 (UTC -12:00)

Block or report othnielObasi

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
othnielObasi/README.md

About Me

I build transparent, scalable AI systems focused on efficiency and governance, specializing in model optimization (distillation & quantization), domain-adapted tokenization, and knowledge-graph–powered reasoning.

My work includes training custom tokenizers, resizing model embeddings to align with bespoke vocabularies, and fine-tuning lightweight LLMs for specialized domains.

Across multiple projects, I’ve delivered LLM systems with grounding, explainability, observability, and deployment controls, applying consistent safety- and governance-first system design principles across search, conversational AI, and decision-support systems.

🛠️ Tech Stack

  • Core | APIs & Deployment

Python PyTorch HuggingFace | FastAPI Streamlit Docker GCP

  • MLOps & AI Tools | Data & Tools

MLflow Weights&Biases OpenAI LangChain | pandas Supabase Neo4j Git

🚀 Key Projects

  • Domain-Adapted Tokenizer & TinyLLaMA v1 QLoRA Fine-Tuning – Tokenizer Training, Embedding Resize & QLoRA Fine-Tuning:
    Trained a BPE tokenizer on a 500 M-token biomedical corpus; resized TinyLLaMA v1’s embedding matrix to fit the new vocabulary; applied 4-bit QLoRA with PEFT adapters; fine-tuned for domain QA—achieving a 12% drop in perplexity and a 15% F1 lift.
    Tools: Python, Hugging Face Transformers, Tokenizers, BitsAndBytes, PEFT, PyTorch

  • Model Optimization Pipeline – Distillation, Quantization & Deployment:
    Trained and distilled a CNN for digit recognition; exported to TorchScript and achieved >98% accuracy post-quantization; built a FastAPI backend and Streamlit UI for real-time inference with Supabase logging; containerized with Docker and deployed on GCP Cloud Run; tracked performance with MLflow and W&B.
    Tools: TorchScript, Distillation, Quantization, FastAPI, Streamlit, Supabase, Docker, GCP, MLflow, Weights & Biases

  • Conversational AI with Knowledge Graph:
    Built a chatbot combining GPT, LangChain RAG, and Neo4j to deliver grounded, context-aware responses, improving reliability and traceability in conversational systems.
    Tools: Python, GPT, LangChain, RAG, Neo4j

  • Neural Search:
    Developed a neural search engine using SentenceTransformer embeddings and a vector database to deliver high-accuracy semantic retrieval.
    Tools: Python, NLP, SentenceTransformer, Vector Database

  • Explainable AI (XAI) for Fraud & Decision Support:
    Developed explainability workflows using LIME, SHAP, and DiCE to support investigation, audit, and accountable decision-making; demoed an interactive XAI prototype.
    Tools: Python, LIME, SHAP, DiCE

🏆 Certifications & Training

  • Neo4j Certified Professional & Neo4j Graph Data Science — Neo4j
  • Generative AI Engineering.
  • Applied Data Science Lab — WorldQuant University
  • Red Teaming LLM Applications — Giskard AI & DL
  • Quality & Safety for LLM Applications — WhyLabs & DL
  • OpenTelemetry Observability Lab — Linux Foundation

📈 Research

  • Explainable AI: Developed tools with LIME, SHAP, and DiCE to enhance AI transparency; demoed an XAI prototype on Hugging Face Spaces.
  • Sovereign AI Governance — Independent work exploring policy-aligned, auditable, and safety-aware AI system orchestration.

👥 Let's Connect

  • 💼 Open to AI/ML collaborations and consulting opportunities
  • 📧 Contact via email or LinkedIn · CV

Pinned Loading

  1. othnielObasi othnielObasi Public