MangaVQA and MangaLMM: A Benchmark and Specialized Model for Multimodal Manga Understanding
Paper
β’ 2505.20298 β’ Published
β’ 9
This repository contains the MangaLMM model described in the paper MangaVQA and MangaLMM: A Benchmark and Specialized Model for Multimodal Manga Understanding.
Code: https://github.com/manga109/MangaLMM
Official demo: https://huggingface.co/spaces/yuki-imajuku/MangaLMM-Demo