🤖A list of PaperList of NLP related papers on Github
-
Updated
Jan 19, 2024
🤖A list of PaperList of NLP related papers on Github
Composition of Multimodal Language Models From Scratch
🖼️Latest Papers on Visually(Imagination)-Augmented NLP
Awesome list for attacks on large language models.
Code for the MultipanelVQA benchmark "Muffin or Chihuahua? Challenging Large Vision-Language Models with Multipanel VQA"
Datasets, case studies and benchmarks for extracting structured information from PDFs, HTML files or images, created by the Parsee.ai team. Datasets also on Hugging Face: https://huggingface.co/parsee-ai
A Video Chat Agent with Temporal Prior
MIKO: Multimodal Intention Knowledge Distillation from Large Language Models for Social-Media Commonsense Discover
Open Source + Multilingual MLLM + Fine-tuning + Distillation + More efficient models and learning + ?
This repository includes the official implementation of our paper "Sight Beyond Text: Multi-Modal Training Enhances LLMs in Truthfulness and Ethics"
Evaluation framework for paper "VisualWebBench: How Far Have Multimodal LLMs Evolved in Web Page Understanding and Grounding?"
Unified Multi-modal IAA Baseline and Benchmark
mPLUG-HalOwl: Multimodal Hallucination Evaluation and Mitigating
A collection of visual instruction tuning datasets.
Official code for Paper "Mantis: Multi-Image Instruction Tuning"
Add a description, image, and links to the mllm topic page so that developers can more easily learn about it.
To associate your repository with the mllm topic, visit your repo's landing page and select "manage topics."