About Me

I'm a AI Engineer and Software Developer who loves the thrill of redefining what it possible. I have an Associate's Degree from Tusculum University and I am currently enrolled there working toward my bachelor's. Currently, I work as an AI Engineer building and training Large Language Models and other types of AI models from next-token prediction to span corruption denoising to embedding and classification. I also do app development, web development, and other things from time to time – like maintenance on this website (which I initially built at 15) though not nearly as much as I used to.

I've been programming for over 7 years – since I was 10. I officially published my first project when I was 11 on Replit. Since then I've tried out building all sorts of things from search engines to compilers to image formats to websites to apps to games, and so much more. Many of those past projects are listed here on this website.

Now that I've been working professionally for a few years, I haven't put out as many public published projects – but that doesn't mean I haven't been building. AI is my passion now and I love the experience of building, curating, and processing training data for the models I build. I also love to design new architectures and implement things I find on ArXiv. Here you can find a tiny portion of my past projects. If you want information on my experience and education, I encourage you to check out my LinkedIn.

Recent Projects

Expressionizer

A Python library for symbolic math expression building, simplification, and step-by-step evaluation. Renders explanations in plain text and LaTeX, supports derivatives, integrals, and multivariate calculus, and ships with localization packs for 7 languages. I primarily use it to procedurally generate math training data with full solution traces.

DAT-Byte-Small

A 200M-parameter byte-level Differential Attention Transformer I trained from scratch on an RTX 5090. Decoder-only, 28 layers, 768 hidden, RoPE, pre-LN, and a 259-token byte-level vocabulary. Trained on Gutenberg English, OpenDiscord, and public-domain Bible translations for 31,200 steps across an estimated 5–10B tokens.

MMLU-NGRAM

A robustness benchmark I built and published that rewrites MMLU questions into character n-grams (n=1 to n=4) to test how well LLMs read unconventional, hard-to-parse inputs. Benchmarked Grok-3-mini, GPT-4.1, Gemini 2.5 Flash, Qwen3-235B, and others. Grok-3-mini led with 87%+ across all n; most models dropped 20–30 points from the original.

See the full archive →