49 results found Sort:
- Filter by Primary Language:
- Python (35)
- Jupyter Notebook (7)
- JavaScript (1)
- Lua (1)
- MATLAB (1)
- +
A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)
Created
2018-06-27
1,097 commits to main branch, last one 7 days ago
InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, ...
Created
2023-05-08
260 commits to main branch, last one 6 months ago
北京航空航天大学大数据高精尖中心自然语言处理研究团队开展了智能问答的研究与应用总结。包括基于知识图谱的问答(KBQA),基于文本的问答系统(TextQA),基于表格的问答系统(TableQA)、基于视觉的问答系统(VisualQA)和机器阅读理解(MRC)等,每类任务分别对学术界和工业界进行了相关总结。
Created
2020-04-28
215 commits to master branch, last one about a year ago
Bottom-up attention model for image captioning and VQA, based on Faster R-CNN and Visual Genome
Created
2017-05-26
55 commits to master branch, last one 3 years ago
The implementation of "Prismer: A Vision-Language Model with Multi-Task Experts".
Created
2023-03-02
36 commits to main branch, last one 4 months ago
Oscar and VinVL
Created
2020-05-14
28 commits to master branch, last one 9 months ago
An efficient PyTorch implementation of the winning entry of the 2017 VQA Challenge.
Created
2017-12-16
24 commits to master branch, last one 4 years ago
[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-bas...
Created
2021-03-23
77 commits to main branch, last one about a year ago
[CVPR 2021 Best Student Paper Honorable Mention, Oral] Official PyTorch code for ClipBERT, an efficient framework for end-to-end learning on image-text and video-text tasks.
Created
2021-02-10
14 commits to main branch, last one about a year ago
A curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area.
Created
2019-03-03
36 commits to master branch, last one about a year ago
Open-source evaluation toolkit of large vision-language models (LVLMs), support GPT-4v, Gemini, QwenVLPlus, 50+ HF models, 20+ benchmarks
Created
2023-12-01
488 commits to main branch, last one 24 hours ago
Implementation for the paper "Compositional Attention Networks for Machine Reasoning" (Hudson and Manning, ICLR 2018)
Created
2018-04-14
71 commits to master branch, last one 3 years ago
Chatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark vision-language models side-by-side while providing images as inputs. Supports MiniGPT-4, LLaMA-Adapter V2, LLaVA, BLIP...
Created
2023-05-10
86 commits to main branch, last one about a month ago
PyTorch implementation of "Transparency by Design: Closing the Gap Between Performance and Interpretability in Visual Reasoning"
Created
2018-03-13
34 commits to master branch, last one 5 years ago
A lightweight, scalable, and general framework for visual question answering research
Created
2019-07-04
254 commits to master branch, last one 2 years ago
Strong baseline for visual question answering
Created
2017-07-30
18 commits to master branch, last one about a year ago
[ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning
Created
2023-06-15
366 commits to main branch, last one 2 months ago
mPLUG-2: A Modularized Multi-modal Foundation Model Across Text, Image and Video (ICML 2023)
Created
2023-05-22
4 commits to main branch, last one 10 months ago
Multimodal Question Answering in the Medical Domain: A summary of Existing Datasets and Systems
Created
2020-04-23
12 commits to master branch, last one 7 months ago
OmniFusion — a multimodal model to communicate using text and images
Created
2023-11-20
80 commits to main branch, last one about a month ago
Research Code for ICCV 2019 paper "Relation-aware Graph Attention Network for Visual Question Answering"
Created
2019-10-07
14 commits to master branch, last one 3 years ago
[NeurIPS 2022] Zero-Shot Video Question Answering via Frozen Bidirectional Language Models
Created
2022-09-25
16 commits to main branch, last one 8 months ago
[IEEE TIP'2021] "UGC-VQA: Benchmarking Blind Video Quality Assessment for User Generated Content", Zhengzhong Tu, Yilin Wang, Neil Birkbeck, Balu Adsumilli, Alan C. Bovik
Created
2020-05-18
41 commits to master branch, last one 2 years ago
[ICCV 2021 Oral + TPAMI] Just Ask: Learning to Answer Questions from Millions of Narrated Videos
Created
2021-03-22
47 commits to main branch, last one 8 months ago
[CVPR 2021] Counterfactual VQA: A Cause-Effect Look at Language Bias
Created
2021-03-02
9 commits to master branch, last one 2 years ago
Code release for ICLR 2023 paper: SlotFormer on object-centric dynamics models
Created
2023-01-23
35 commits to master branch, last one 8 months ago
Using pretrained encoder and language models to generate captions from multimedia inputs.
Created
2022-01-23
507 commits to main branch, last one about a year ago
mPLUG: Effective and Efficient Vision-Language Learning by Cross-modal Skip-connections. (EMNLP 2022)
Created
2023-05-08
4 commits to main branch, last one about a year ago
Visual Question Answering in the Medical Domain VQA-Med 2019
Created
2019-05-22
25 commits to master branch, last one 4 months ago
Counterfactual Samples Synthesizing for Robust VQA
Created
2020-02-25
44 commits to master branch, last one about a year ago