Hello! I am wondering, since this would be a very interesting use case and there is more than enough training material out there (pretty much every MD file could be rendered, then the image and markdown code could be used for training/finetuning)
however I have pretty much no idea about llava. Do you think this would be feasible to do?
what you are looking for is OCR. then feed the LLM to the markdown
Facebook Nougat OCR model does PDF to Markdown. Also fine tuned versions of it doing PDF to LaTeX. I plan on making a fine tuned version to do PDF to XML later this winter break too!
I’d really like a version of llava that can process comic/manga pages (read the text, say which character is saying what, doing what, in what order. pretty much turn the manga into a novel or something like that).
Anyone know of any project that is going in that direction/working on that?