Faruna Godwin Abuh

Applied AI Safety Engineer ยท Evaluation & Interpretability

Building safety-aware AI systems with focus on model behavior, dataset quality, and real-world deployment

Selected Project Metrics

Exploratory & Demo Deployments ยท Non-Production Research Projects

0
๐Ÿ“Š
Igala-English Parallel
Sentences Collected
0
๐Ÿ”ฌ
Safety & NLP Tools
Deployed
0+
๐ŸŽฏ
Vulnerability Categories
Tested
Focus on interpretability, robustness, and dataset quality.
All tools built for research exploration and learning purposes.
๐Ÿ›ก๏ธ LLM Red-Teaming Framework
๐Ÿ”ฌ Mechanistic Interpretability
๐ŸŒ First Igala NMT System
โšก Custom Igala GPT from Scratch

About / Research Philosophy

I focus on applied AI safety โ€” how models behave in real systems, how datasets shape outcomes, and how deployment context introduces risk. My work centers on evaluation, interpretability, and low-resource NLP, where safety issues often appear earlier and more visibly.

Rather than optimizing for scale or performance alone, I prioritize transparency, robustness, and understanding failure modes before deployment. I'm especially interested in how AI safety frameworks apply outside well-resourced settings and how tooling can support safer adoption globally.

Featured Projects

Building AI for Real-World Impact

Production workbench for systematically detecting safety decay in frontier LLMs (Gemini, GPT-4, Claude) when prompted in Yoruba, Hausa, Igbo, and Igala. Features dual-probe comparison, mechanistic visualization, and session analytics for demonstrating 2-4x higher bypass rates in low-resource contexts.

โœ“Side-by-side English baseline vs target language analysis
โœ“Automatic loophole detection with session statistics
โœ“18 attack vectors across 4 African languages
โœ“Mechanistic visualizations: activation smearing & centroid drift
โœ“Exportable JSON logs for offline analysis
โœ“Google Translate verification integration
StreamlitPlotlyGoogle Gemini APIPythonPandasNumPy
View Full Case Study

Fine-tuned mBERT model for Igala-English translation with quality metrics, uncertainty quantification, and interactive demo. Addresses critical language preservation and accessibility needs.

โœ“3,253 parallel sentences training corpus
โœ“Real-time bidirectional translation
โœ“Confidence scoring for each translation
โœ“First publicly available Igala MT system
PyTorchmBERTTransformersStreamlitHuggingFace
View Full Case Study

Deep dive into mBERT's internal mechanisms during Igala-English translation. Visualizes attention heads, analyzes token alignments, and explores how transformers handle morphologically complex languages.

โœ“Layer-by-layer attention visualization
โœ“Token alignment analysis across 12 attention heads
โœ“Morphological feature tracking
โœ“Interactive attention pattern explorer
PyTorchTransformerLensmBERTStreamlitPlotly
View Full Case Study

A decoder-only transformer trained on Igala corpus, implementing multi-head self-attention, positional encoding, and custom tokenization. Demonstrates deep understanding of transformer architecture.

โœ“Built transformer architecture from scratch (no pretrained models)
โœ“268KB Igala corpus training data
โœ“Custom BPE tokenizer implementation
โœ“Interactive text generation interface
PyTorchCustom TokenizerStreamlitNumPyTransformers
View Full Case Study

An interactive Streamlit application for exploring and auditing low-resource language datasets. The tool provides sentence statistics, token analysis, and linguistic insights to support NLP research and language preservation efforts.

โœ“3,253 field-collected sentences
โœ“First comprehensive Igala NLP dataset
โœ“Interactive data exploration interface
โœ“Designed for expansion to other African languages
PythonPandasStreamlitHuggingFaceNLP
View Full Case Study

Empirical evaluation and mechanistic interpretability analysis of transformer models, with a focus on reliability, calibration, and understanding internal decision mechanisms using GPT-2.

โœ“Direct Logit Attribution analysis on GPT-2
โœ“Selective prediction and abstention methods explored
โœ“Calibration improvements through temperature scaling
โœ“Research-grade evaluation framework
PyTorchTransformerLensGPT-2PythonScikit-learn
View Full Case Study

A full-stack AI chatbot featuring conversation memory, automatic mode detection, real-time logging, and analytics. Built to provide personalized portfolio Q&A experience using Google Gemini API.

โœ“Multi-mode responses (deep-dive, quick, story, default)
โœ“Sub-2 second average response time
โœ“Session persistence across page reloads
โœ“Production logging with analytics dashboard
Next.jsFastAPIGoogle GeminiPythonDocker
View Full Case Study

Project Vision

My long-term goal is to build AI systems that are not only powerful, but also trustworthy and useful in real-world settings. This means designing models that communicate uncertainty, respect local languages, and avoid overconfident or unsafe recommendations.

I'm especially interested in deploying AI where access to expertise is limited, and where responsible design can make a meaningful difference. Whether it's building datasets for underrepresented languages, making models more interpretable, or designing systems that know when to abstain, I believe AI should serve communities, not replace human judgment.

๐ŸŒ

Inclusive NLP

Building for low-resource languages

๐Ÿ”ฌ

AI Safety

Transparency and reliability

๐Ÿค

Real Impact

Serving underserved communities

Let's Connect: farunagodwin01@gmail.com

Open to applied AI safety, evaluation, and research-engineering collaboration. If you're working on model behavior analysis, low-resource NLP, or responsible AI deployment, let's talk.

๐Ÿ’ก Currently exploring: Selective prediction methods, failure mode analysis in low-resource NLP, and production LLM observability