Faruna Godwin Abuh

Applied AI Safety Engineer Β· Evaluation & Interpretability

Building safety-aware AI systems with focus on model behavior, dataset quality, and real-world deployment

Selected Project Metrics

Exploratory & Demo Deployments Β· Non-Production Research Projects

0
πŸ“Š
Igala-English Parallel
Sentences Collected
0
πŸ”¬
Safety & NLP Tools
Deployed
0+
🎯
Vulnerability Categories
Tested
Focus on interpretability, robustness, and dataset quality.
All tools built for research exploration and learning purposes.
πŸ›‘οΈ LLM Red-Teaming Framework
πŸ”¬ Mechanistic Interpretability
🌍 First Igala NMT System
⚑ Custom Igala GPT from Scratch

About / Research Philosophy

I focus on applied AI safety β€” how models behave in real systems, how datasets shape outcomes, and how deployment context introduces risk. My work centers on evaluation, interpretability, and low-resource NLP, where safety issues often appear earlier and more visibly.

Rather than optimizing for scale or performance alone, I prioritize transparency, robustness, and understanding failure modes before deployment. I'm especially interested in how AI safety frameworks apply outside well-resourced settings and how tooling can support safer adoption globally.

Featured Projects

Building AI for Real-World Impact

Systematic red-teaming framework for evaluating LLM robustness against prompt injection, jailbreaks, and unsafe outputs. Implements multiple attack strategies with automated safety scoring.

βœ“Tests 5+ vulnerability categories (injection, bias, toxicity)
βœ“Automated attack generation and evaluation
βœ“Comprehensive safety scoring framework
βœ“Production-ready testing pipeline
PythonTransformersStreamlitOpenAISafety Metrics
View Full Case Study

Fine-tuned mBERT model for Igala-English translation with quality metrics, uncertainty quantification, and interactive demo. Addresses critical language preservation and accessibility needs.

βœ“3,253 parallel sentences training corpus
βœ“Real-time bidirectional translation
βœ“Confidence scoring for each translation
βœ“First publicly available Igala MT system
PyTorchmBERTTransformersStreamlitHuggingFace
View Full Case Study

Deep dive into mBERT's internal mechanisms during Igala-English translation. Visualizes attention heads, analyzes token alignments, and explores how transformers handle morphologically complex languages.

βœ“Layer-by-layer attention visualization
βœ“Token alignment analysis across 12 attention heads
βœ“Morphological feature tracking
βœ“Interactive attention pattern explorer
PyTorchTransformerLensmBERTStreamlitPlotly
View Full Case Study

A decoder-only transformer trained on Igala corpus, implementing multi-head self-attention, positional encoding, and custom tokenization. Demonstrates deep understanding of transformer architecture.

βœ“Built transformer architecture from scratch (no pretrained models)
βœ“268KB Igala corpus training data
βœ“Custom BPE tokenizer implementation
βœ“Interactive text generation interface
PyTorchCustom TokenizerStreamlitNumPyTransformers
View Full Case Study

An interactive Streamlit application for exploring and auditing low-resource language datasets. The tool provides sentence statistics, token analysis, and linguistic insights to support NLP research and language preservation efforts.

βœ“3,253 field-collected sentences
βœ“First comprehensive Igala NLP dataset
βœ“Interactive data exploration interface
βœ“Designed for expansion to other African languages
PythonPandasStreamlitHuggingFaceNLP
View Full Case Study

Empirical evaluation and mechanistic interpretability analysis of transformer models, with a focus on reliability, calibration, and understanding internal decision mechanisms using GPT-2.

βœ“Direct Logit Attribution analysis on GPT-2
βœ“Selective prediction and abstention methods explored
βœ“Calibration improvements through temperature scaling
βœ“Research-grade evaluation framework
PyTorchTransformerLensGPT-2PythonScikit-learn
View Full Case Study

A full-stack AI chatbot featuring conversation memory, automatic mode detection, real-time logging, and analytics. Built to provide personalized portfolio Q&A experience using Google Gemini API.

βœ“Multi-mode responses (deep-dive, quick, story, default)
βœ“Sub-2 second average response time
βœ“Session persistence across page reloads
βœ“Production logging with analytics dashboard
Next.jsFastAPIGoogle GeminiPythonDocker
View Full Case Study

Project Vision

My long-term goal is to build AI systems that are not only powerful, but also trustworthy and useful in real-world settings. This means designing models that communicate uncertainty, respect local languages, and avoid overconfident or unsafe recommendations.

I'm especially interested in deploying AI where access to expertise is limited, and where responsible design can make a meaningful difference. Whether it's building datasets for underrepresented languages, making models more interpretable, or designing systems that know when to abstainβ€”I believe AI should serve communities, not replace human judgment.

🌍

Inclusive NLP

Building for low-resource languages

πŸ”¬

AI Safety

Transparency and reliability

🀝

Real Impact

Serving underserved communities

Let's Connect

Open to applied AI safety, evaluation, and research-engineering collaboration. If you're working on model behavior analysis, low-resource NLP, or responsible AI deploymentβ€”let's talk.

πŸ’‘ Currently exploring: Selective prediction methods, failure mode analysis in low-resource NLP, and production LLM observability