DatBench: Discriminative, Faithful, and Efficient VLM Evaluations
A benchmark and evaluation framework for vision-language models built around discriminative tasks, faithful scoring, and practical efficiency.

DatBench leads the page as the current benchmark story, followed by earlier papers in multimodal representation learning and medical imaging.
Current work on making vision-language model evaluation more discriminative, more faithful, and practical enough to run in real research loops.
A benchmark and evaluation framework for vision-language models built around discriminative tasks, faithful scoring, and practical efficiency.

Earlier work is presented as a quieter reading list, with explicit links out to the paper, code, or publisher page where available.
Transformer-based multimodal re-identification focused on stronger fusion and cleaner representation learning.