Publications

Papers and published research

DatBench leads the page as the current benchmark story, followed by earlier papers in multimodal representation learning and medical imaging.

Featured research

Current work on making vision-language model evaluation more discriminative, more faithful, and practical enough to run in real research loops.

Research archive

Earlier papers

Earlier work is presented as a quieter reading list, with explicit links out to the paper, code, or publisher page where available.

UniCat
arXiv2023

UniCat: Crafting a Stronger Fusion Baseline for Multimodal Re-Identification

J. Crawford, H. Yin, L. McDermott, D. Cummings

A stronger multimodal fusion baseline for re-identification, paired with an open-source release.

GraFT
arXiv2023

GraFT: Gradual Fusion Transformer for Multimodal Re-Identification

H. Yin, J. Li, E. Schiller, L. McDermott, D. Cummings

Transformer-based multimodal re-identification focused on stronger fusion and cleaner representation learning.

SpecReFlow
Journal of Medical Imaging2024

SpecReFlow: an algorithm for specular reflection restoration using flow-guided video completion

H. Yin, R. Eimen, D. Moyer, A. K. Bowden

Reflection-aware medical video restoration combining optical flow with video completion.