My research focuses on bridging the gap between academic rigor and production-scale machine learning. The common thread across my work is applying robust representation learning to complex, unstructured domains. Whether I am extracting structured data from messy business documents, forecasting chaotic physical systems, or probing the vulnerabilities of language models, my goal is to build systems that are both mathematically grounded and practically useful.

My primary focus areas include:

Selected Publications

If you are new to my work, I recommend starting here:

  1. GutenOCR: A Grounded Vision-Language Front-End for Documents (2026): A family of VLMs fine-tuned for precise text transcription and explicit geometric grounding.
  2. Page Stream Segmentation with LLMs: Challenges and Applications in Insurance Document Automation (COLING 2025): Real-world application and calibration analysis of LLMs for document automation.
  3. Deconstructing recurrence, attention, and gating (2024): An ablation study on sequence models for forecasting high-dimensional chaotic systems.

Below you find my published work, preprints, and open-source datasets. For citation metrics, please see my Google Scholar profile.