Benchmark and evaluation systems for VLM research
Built evaluation paths that keep model comparisons useful for curation decisions and model iteration.
The current emphasis is research engineering for multimodal training and evaluation, with a bias toward faster iteration.
At Datology, I build the systems that make VLM training and evaluation more reliable in practice.
Built evaluation paths that keep model comparisons useful for curation decisions and model iteration.
Built ingestion and export paths that make large multimodal corpora easier to train on and inspect.
Added vLLM eval support and hardened multi-node launch plus checkpoint behavior for faster experimental turnover.
Selected public projects presented as quieter portfolio entries rather than product tiles.