René's URL Explorer Experiment


Title: GitHub - billiecn/visual-question-answering: A curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area.

Open Graph Title: GitHub - billiecn/visual-question-answering: A curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area.

X Title: GitHub - billiecn/visual-question-answering: A curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area.

Description: A curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area. - billiecn/visual-question-answering

Open Graph Description: A curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area. - billiecn/visual-question...

X Description: A curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area. - billiecn/visual-question...

Mail addresses
leungjokie@gmail.com

Opengraph URL: https://github.com/billiecn/visual-question-answering

X: @github

direct link

Domain: patch-diff.githubusercontent.com

route-pattern/:user_id/:repository
route-controllerfiles
route-actiondisambiguate
fetch-noncev2:70c27a61-2ba6-fad8-4e4c-3f64996377a5
current-catalog-service-hashf3abb0cc802f3d7b95fc8762b94bdcb13bf39634c40c357301c4aa1d67a256fb
request-id97BE:2464D3:14F91E9:1B5B40B:6992AE6F
html-safe-noncec08661e39ed829d8db93ed81fd6a16caa1866f4e35ca3ca15159733bbc7b3e7f
visitor-payloadeyJyZWZlcnJlciI6IiIsInJlcXVlc3RfaWQiOiI5N0JFOjI0NjREMzoxNEY5MUU5OjFCNUI0MEI6Njk5MkFFNkYiLCJ2aXNpdG9yX2lkIjoiMTU1ODE0ODgyNzQ4NzMxNzYxNiIsInJlZ2lvbl9lZGdlIjoiaWFkIiwicmVnaW9uX3JlbmRlciI6ImlhZCJ9
visitor-hmacb6786d76438830938ea889e00fdf1d69cfe4b83e5eb9b073977b48276e4c66ef
hovercard-subject-tagrepository:494461472
github-keyboard-shortcutsrepository,copilot
google-site-verificationApib7-x98H0j5cPqHWwSMm6dNU4GmODRoqxLiDzdx9I
octolytics-urlhttps://collector.github.com/github/collect
analytics-location//
fb:app_id1401488693436528
apple-itunes-appapp-id=1477376905, app-argument=https://github.com/billiecn/visual-question-answering
twitter:imagehttps://opengraph.githubassets.com/fdb32c313a442b78caa44eb188d521a6ddde9a6e09f04924557c337680bfd40c/billiecn/visual-question-answering
twitter:cardsummary_large_image
og:imagehttps://opengraph.githubassets.com/fdb32c313a442b78caa44eb188d521a6ddde9a6e09f04924557c337680bfd40c/billiecn/visual-question-answering
og:image:altA curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area. - billiecn/visual-question...
og:image:width1200
og:image:height600
og:site_nameGitHub
og:typeobject
hostnamegithub.com
expected-hostnamegithub.com
None42c603b9d642c4a9065a51770f75e5e27132fef0e858607f5c9cb7e422831a7b
turbo-cache-controlno-preview
go-importgithub.com/billiecn/visual-question-answering git https://github.com/billiecn/visual-question-answering.git
octolytics-dimension-user_id75419179
octolytics-dimension-user_loginbilliecn
octolytics-dimension-repository_id494461472
octolytics-dimension-repository_nwobilliecn/visual-question-answering
octolytics-dimension-repository_publictrue
octolytics-dimension-repository_is_forktrue
octolytics-dimension-repository_parent_id173578866
octolytics-dimension-repository_parent_nwojokieleung/awesome-visual-question-answering
octolytics-dimension-repository_network_root_id173578866
octolytics-dimension-repository_network_root_nwojokieleung/awesome-visual-question-answering
turbo-body-classeslogged-out env-production page-responsive
disable-turbofalse
browser-stats-urlhttps://api.github.com/_private/browser/stats
browser-errors-urlhttps://api.github.com/_private/browser/errors
release848bc6032dcc93a9a7301dcc3f379a72ba13b96e
ui-targetfull
theme-color#1e2327
color-schemelight dark

Links:

Skip to contenthttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#start-of-content
https://patch-diff.githubusercontent.com/
Sign in https://patch-diff.githubusercontent.com/login?return_to=https%3A%2F%2Fgithub.com%2Fbilliecn%2Fvisual-question-answering
GitHub CopilotWrite better code with AIhttps://github.com/features/copilot
GitHub SparkBuild and deploy intelligent appshttps://github.com/features/spark
GitHub ModelsManage and compare promptshttps://github.com/features/models
MCP RegistryNewIntegrate external toolshttps://github.com/mcp
ActionsAutomate any workflowhttps://github.com/features/actions
CodespacesInstant dev environmentshttps://github.com/features/codespaces
IssuesPlan and track workhttps://github.com/features/issues
Code ReviewManage code changeshttps://github.com/features/code-review
GitHub Advanced SecurityFind and fix vulnerabilitieshttps://github.com/security/advanced-security
Code securitySecure your code as you buildhttps://github.com/security/advanced-security/code-security
Secret protectionStop leaks before they starthttps://github.com/security/advanced-security/secret-protection
Why GitHubhttps://github.com/why-github
Documentationhttps://docs.github.com
Bloghttps://github.blog
Changeloghttps://github.blog/changelog
Marketplacehttps://github.com/marketplace
View all featureshttps://github.com/features
Enterpriseshttps://github.com/enterprise
Small and medium teamshttps://github.com/team
Startupshttps://github.com/enterprise/startups
Nonprofitshttps://github.com/solutions/industry/nonprofits
App Modernizationhttps://github.com/solutions/use-case/app-modernization
DevSecOpshttps://github.com/solutions/use-case/devsecops
DevOpshttps://github.com/solutions/use-case/devops
CI/CDhttps://github.com/solutions/use-case/ci-cd
View all use caseshttps://github.com/solutions/use-case
Healthcarehttps://github.com/solutions/industry/healthcare
Financial serviceshttps://github.com/solutions/industry/financial-services
Manufacturinghttps://github.com/solutions/industry/manufacturing
Governmenthttps://github.com/solutions/industry/government
View all industrieshttps://github.com/solutions/industry
View all solutionshttps://github.com/solutions
AIhttps://github.com/resources/articles?topic=ai
Software Developmenthttps://github.com/resources/articles?topic=software-development
DevOpshttps://github.com/resources/articles?topic=devops
Securityhttps://github.com/resources/articles?topic=security
View all topicshttps://github.com/resources/articles
Customer storieshttps://github.com/customer-stories
Events & webinarshttps://github.com/resources/events
Ebooks & reportshttps://github.com/resources/whitepapers
Business insightshttps://github.com/solutions/executive-insights
GitHub Skillshttps://skills.github.com
Documentationhttps://docs.github.com
Customer supporthttps://support.github.com
Community forumhttps://github.com/orgs/community/discussions
Trust centerhttps://github.com/trust-center
Partnershttps://github.com/partners
GitHub SponsorsFund open source developershttps://github.com/sponsors
Security Labhttps://securitylab.github.com
Maintainer Communityhttps://maintainers.github.com
Acceleratorhttps://github.com/accelerator
Archive Programhttps://archiveprogram.github.com
Topicshttps://github.com/topics
Trendinghttps://github.com/trending
Collectionshttps://github.com/collections
Enterprise platformAI-powered developer platformhttps://github.com/enterprise
GitHub Advanced SecurityEnterprise-grade security featureshttps://github.com/security/advanced-security
Copilot for BusinessEnterprise-grade AI featureshttps://github.com/features/copilot/copilot-business
Premium SupportEnterprise-grade 24/7 supporthttps://github.com/premium-support
Pricinghttps://github.com/pricing
Search syntax tipshttps://docs.github.com/search-github/github-code-search/understanding-github-code-search-syntax
documentationhttps://docs.github.com/search-github/github-code-search/understanding-github-code-search-syntax
Sign in https://patch-diff.githubusercontent.com/login?return_to=https%3A%2F%2Fgithub.com%2Fbilliecn%2Fvisual-question-answering
Sign up https://patch-diff.githubusercontent.com/signup?ref_cta=Sign+up&ref_loc=header+logged+out&ref_page=%2F%3Cuser-name%3E%2F%3Crepo-name%3E&source=header-repo&source_repo=billiecn%2Fvisual-question-answering
Reloadhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering
Reloadhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering
Reloadhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering
billiecn https://patch-diff.githubusercontent.com/billiecn
visual-question-answeringhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering
jokieleung/awesome-visual-question-answeringhttps://patch-diff.githubusercontent.com/jokieleung/awesome-visual-question-answering
Notifications https://patch-diff.githubusercontent.com/login?return_to=%2Fbilliecn%2Fvisual-question-answering
Fork 1 https://patch-diff.githubusercontent.com/login?return_to=%2Fbilliecn%2Fvisual-question-answering
Star 1 https://patch-diff.githubusercontent.com/login?return_to=%2Fbilliecn%2Fvisual-question-answering
1 star https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/stargazers
95 forks https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/forks
Branches https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/branches
Tags https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/tags
Activity https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/activity
Star https://patch-diff.githubusercontent.com/login?return_to=%2Fbilliecn%2Fvisual-question-answering
Notifications https://patch-diff.githubusercontent.com/login?return_to=%2Fbilliecn%2Fvisual-question-answering
Code https://patch-diff.githubusercontent.com/billiecn/visual-question-answering
Pull requests 0 https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/pulls
Actions https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/actions
Projects 0 https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/projects
Security 0 https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/security
Insights https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/pulse
Code https://patch-diff.githubusercontent.com/billiecn/visual-question-answering
Pull requests https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/pulls
Actions https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/actions
Projects https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/projects
Security https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/security
Insights https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/pulse
Brancheshttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/branches
Tagshttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/tags
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/branches
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/tags
28 Commitshttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/commits/master/
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering/commits/master/
README.mdhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master/README.md
README.mdhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master/README.md
contributing.mdhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master/contributing.md
contributing.mdhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master/contributing.md
READMEhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering
Contributinghttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering
https://awesome.re
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#awesome-visual-question-answering
https://camo.githubusercontent.com/1131548cf666e1150ebd2a52f44776d539f06324/68747470733a2f2f63646e2e7261776769742e636f6d2f73696e647265736f726875732f617765736f6d652f6d61737465722f6d656469612f6c6f676f2e737667
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#contributing
pull requestshttps://github.com/jokieleung/awesome-visual-question-answering/pulls
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#change-log
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#table-of-contents
Contributinghttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#contributing
Change Loghttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#change-log
Table of Contentshttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#table-of-contents
Papershttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#papers
Surveyhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#survey
2022https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2022
ACL 2022https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#ACL-2022
CVPR 2022https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#CVPR-2022
AAAI 2022https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#AAAI-2022
2021https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2021
NeurIPS 2021https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#NeurIPS-2021
EMNLP 2021https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#EMNLP-2021
ICCV 2021https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#ICCV-2021
ACL 2021https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#ACL-2021
SIGIR 2021https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#SIGIR-2021
CVPR 2021https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#CVPR-2021
ICLR 2021https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#ICLR-2021
NAACL-HLT 2021https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#NAACL-HLT-2021
AAAI 2021https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#AAAI-2021
2020https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2020
EMNLP 2020https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#EMNLP-2020
NeurIPS 2020https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#NeurIPS-2020
ECCV 2020https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#ECCV-2020
CVPR 2020https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#CVPR-2020
ACL 2020https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#ACL-2020
WACV 2020https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#WACV-2020
AAAI 2020https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#AAAI-2020
2019https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2019
ACL 2019https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#ACL-2019
ICCV 2019https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#ICCV-2019
NeurIPS 2019https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#NeurIPS-2019
CVPR 2019https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#cvpr-2019
AAAI 2019https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#aaai-2019
OTHERhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#other
2018https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2018
NIPS 2018https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#nips-2018
AAAI 2018https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#aaai-2018
IJCAI 2018https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#ijcai-2018
CVPR 2018https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#cvpr-2018
ACM MM 2018https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#acm-mm-2018
ECCV 2018https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#eccv-2018
OTHERhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#other
2017-2015https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2017-2015
OTHERhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#other-1
ICCV 2017https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#iccv-2017
VQA Challenge Leaderboardhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#vqa-challenge-leaderboard
test-std 2018https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#test-std-2018
test-std 2017https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#test-std-2017
Licenseshttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#licenses
Reference and Acknowledgementhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering#reference-and-acknowledgement
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#papers
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#survey
Visual question answering: Datasets, algorithms, and future challengeshttps://arxiv.org/abs/1610.01465
Visual question answering: A survey of methods and datasetshttps://arxiv.org/abs/1607.05910
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2022
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#acl-2022
xGQA: Cross-Lingual Visual Question Answeringhttps://arxiv.org/abs/2109.06082
[data]https://github.com/Adapter-Hub/xGQA
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#cvpr-2022
SimVQA: Exploring Simulated Environments for Visual Question Answeringhttps://arxiv.org/abs/2203.17219
[code]https://www.cs.rice.edu/~pc51/simvqa/
A Thousand Words Are Worth More Than a Picture: Natural Language-Centric Outside-Knowledge Visual Question Answeringhttps://arxiv.org/abs/2201.05299
SwapMix: Diagnosing and Regularizing the Over-reliance on Visual Context in Visual Question Answeringhttps://arxiv.org/abs/2204.02285
[code]https://github.com/vipulgupta1011/swapmix
Dual-Key Multimodal Backdoors for Visual Question Answeringhttps://arxiv.org/abs/2112.07668
[code]https://github.com/SRI-CSL/TrinityMultimodalTrojAI
MuKEA: Multimodal Knowledge Extraction and Accumulation for Knowledge-based Visual Question Answeringhttps://arxiv.org/abs/2203.09138
[code]https://github.com/AndersonStra/MuKEA
Grounding Answers for Visual Questions Asked by Visually Impaired Peoplehttps://arxiv.org/abs/2202.01993
[page]https://vizwiz.org/tasks-and-datasets/answer-grounding-for-vqa/
Maintaining Reasoning Consistency in Compositional Visual Question Answeringhttps://jingchenchen.github.io/files/papers/2022/CVPR_DLR.pdf
[code]https://github.com/jingchenchen/ReasoningConsistency-VQA
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#aaai-2022
Dynamic Key-value Memory Enhanced Multi-step Graph Reasoning for Knowledge-based Visual Question Answeringhttps://arxiv.org/abs/2203.02985
[code]https://github.com/Mingxiao-Li/DMMGR
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2021
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#neurips-2021
Human-Adversarial Visual Question Answeringhttps://arxiv.org/abs/2106.02280
[code]https://adversarialvqa.org/
Debiased Visual Question Answering from Feature and Sample Perspectiveshttps://openreview.net/pdf?id=Z4ry59PVMq8
[code]https://github.com/Zhiquan-Wen/D-VQA
Learning to Generate Visual Questions with Noisy Supervisionhttps://openreview.net/pdf?id=LMotP3zsq_d
[code]https://github.com/AlanSwift/DH-GAN
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#emnlp-2021
Discovering the Unknown Knowns: Turning Implicit Knowledge in the Dataset into Explicit Training Examples for Visual Question Answeringhttps://aclanthology.org/2021.emnlp-main.512/
Beyond Accuracy: A Consolidated Tool for Visual Question Answering Benchmarkinghttps://aclanthology.org/2021.emnlp-demo.14/
[code]https://github.com/patilli/vqa_benchmarking
Diversity and Consistency: Exploring Visual Question-Answer Pair Generationhttps://aclanthology.org/2021.findings-emnlp.91/
Towards Developing a Multilingual and Code-Mixed Visual Question Answering System by Knowledge Distillationhttps://aclanthology.org/2021.findings-emnlp.151/
MIRTT: Learning Multimodal Interaction Representations from Trilinear Transformers for Visual Question Answeringhttps://aclanthology.org/2021.findings-emnlp.196/
[code]https://github.com/iigroup/mirtt
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#iccv-2021
Just Ask: Learning To Answer Questions From Millions of Narrated Videoshttps://openaccess.thecvf.com/content/ICCV2021/html/Yang_Just_Ask_Learning_To_Answer_Questions_From_Millions_of_Narrated_ICCV_2021_paper.html
Env-QA: A Video Question Answering Benchmark for Comprehensive Understanding of Dynamic Environmentshttps://openaccess.thecvf.com/content/ICCV2021/html/Gao_Env-QA_A_Video_Question_Answering_Benchmark_for_Comprehensive_Understanding_of_ICCV_2021_paper.html
On The Hidden Treasure of Dialog in Video Question Answeringhttps://openaccess.thecvf.com/content/ICCV2021/html/Engin_On_the_Hidden_Treasure_of_Dialog_in_Video_Question_Answering_ICCV_2021_paper.html
Unshuffling Data for Improved Generalization in Visual Question Answeringhttps://openaccess.thecvf.com/content/ICCV2021/html/Teney_Unshuffling_Data_for_Improved_Generalization_in_Visual_Question_Answering_ICCV_2021_paper.html
TRAR: Routing the Attention Spans in Transformer for Visual Question Answeringhttps://openaccess.thecvf.com/content/ICCV2021/html/Zhou_TRAR_Routing_the_Attention_Spans_in_Transformer_for_Visual_Question_ICCV_2021_paper.html
Greedy Gradient Ensemble for Robust Visual Question Answeringhttps://openaccess.thecvf.com/content/ICCV2021/html/Han_Greedy_Gradient_Ensemble_for_Robust_Visual_Question_Answering_ICCV_2021_paper.html
Pano-AVQA: Grounded Audio-Visual Question Answering on 360deg Videoshttps://openaccess.thecvf.com/content/ICCV2021/html/Yun_Pano-AVQA_Grounded_Audio-Visual_Question_Answering_on_360deg_Videos_ICCV_2021_paper.html
Weakly Supervised Relative Spatial Reasoning for Visual Question Answeringhttps://openaccess.thecvf.com/content/ICCV2021/html/Banerjee_Weakly_Supervised_Relative_Spatial_Reasoning_for_Visual_Question_Answering_ICCV_2021_paper.html
Linguistically Routing Capsule Network for Out-of-Distribution Visual Question Answeringhttps://openaccess.thecvf.com/content/ICCV2021/html/Cao_Linguistically_Routing_Capsule_Network_for_Out-of-Distribution_Visual_Question_Answering_ICCV_2021_paper.html
Beyond Question-Based Biases: Assessing Multimodal Shortcut Learning in Visual Question Answeringhttps://openaccess.thecvf.com/content/ICCV2021/html/Dancette_Beyond_Question-Based_Biases_Assessing_Multimodal_Shortcut_Learning_in_Visual_Question_ICCV_2021_paper.html
Auto-Parsing Network for Image Captioning and Visual Question Answeringhttps://openaccess.thecvf.com/content/ICCV2021/html/Yang_Auto-Parsing_Network_for_Image_Captioning_and_Visual_Question_Answering_ICCV_2021_paper.html
Unified Questioner Transformer for Descriptive Question Generation in Goal-Oriented Visual Dialoguehttps://openaccess.thecvf.com/content/ICCV2021/html/Matsumori_Unified_Questioner_Transformer_for_Descriptive_Question_Generation_in_Goal-Oriented_Visual_ICCV_2021_paper.html
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#acl-2021
Check It Again:Progressive Visual Question Answering via Visual Entailmenthttps://aclanthology.org/2021.acl-long.317.pdf
[code]https://github.com/PhoebusSi/SAR
Mind Your Outliers! Investigating the Negative Impact of Outliers on Active Learning for Visual Question Answeringhttps://aclanthology.org/2021.acl-long.564.pdf
[code]https://github.com/siddk/vqa-outliers
In Factuality: Efficient Integration of Relevant Facts for Visual Question Answeringhttps://aclanthology.org/2021.acl-short.60.pdf
Towards Visual Question Answering on Pathology Imageshttps://aclanthology.org/2021.acl-short.90.pdf
[code]https://github.com/UCSD-AI4H/PathVQA
Are VQA Systems RAD? Measuring Robustness to Augmented Data with Focused Interventionshttps://aclanthology.org/2021.acl-short.10.pdf
[code]https://danrosenberg.github.io/rad-measure/
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#sigir-2021
LPF: A Language-Prior Feedback Objective Function for De-biased Visual Question Answeringhttps://arxiv.org/abs/2105.14300
[code]https://github.com/jokieleung/LPF-VQA
Passage Retrieval for Outside-Knowledge Visual Question Answeringhttps://arxiv.org/abs/2105.03938
[code]https://github.com/prdwb/okvqa-release
Select, Substitute, Search: A New Benchmark for Knowledge-Augmented Visual Question Answeringhttps://arxiv.org/abs/2103.05568
[code]https://s3vqa.github.io/
Visual Question Rewriting for Increasing Response Ratehttps://arxiv.org/abs/2106.02257
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#cvpr-2021
Separating Skills and Concepts for Novel Visual Question Answeringhttps://openaccess.thecvf.com/content/CVPR2021/html/Whitehead_Separating_Skills_and_Concepts_for_Novel_Visual_Question_Answering_CVPR_2021_paper.html
Roses Are Red, Violets Are Blue... but Should VQA Expect Them To?https://openaccess.thecvf.com/content/CVPR2021/html/Kervadec_Roses_Are_Red_Violets_Are_Blue..._but_Should_VQA_Expect_CVPR_2021_paper.html
[code]https://github.com/gqa-ood/GQA-OOD
Predicting Human Scanpaths in Visual Question Answeringhttps://openaccess.thecvf.com/content/CVPR2021/html/Chen_Predicting_Human_Scanpaths_in_Visual_Question_Answering_CVPR_2021_paper.html
Found a Reason for me? Weakly-supervised Grounded Visual Question Answering using Capsuleshttps://openaccess.thecvf.com/content/CVPR2021/html/Urooj_Found_a_Reason_for_me_Weakly-supervised_Grounded_Visual_Question_Answering_CVPR_2021_paper.html
TAP: Text-Aware Pre-Training for Text-VQA and Text-Captionhttps://openaccess.thecvf.com/content/CVPR2021/html/Yang_TAP_Text-Aware_Pre-Training_for_Text-VQA_and_Text-Caption_CVPR_2021_paper.html
Counterfactual VQA: A Cause-Effect Look at Language Biashttps://openaccess.thecvf.com/content/CVPR2021/html/Niu_Counterfactual_VQA_A_Cause-Effect_Look_at_Language_Bias_CVPR_2021_paper.html
[code]https://github.com/yuleiniu/cfvqa
KRISP: Integrating Implicit and Symbolic Knowledge for Open-Domain Knowledge-Based VQAhttps://openaccess.thecvf.com/content/CVPR2021/html/Marino_KRISP_Integrating_Implicit_and_Symbolic_Knowledge_for_Open-Domain_Knowledge-Based_VQA_CVPR_2021_paper.html
Perception Matters: Detecting Perception Failures of VQA Models Using Metamorphic Testinghttps://openaccess.thecvf.com/content/CVPR2021/html/Yuan_Perception_Matters_Detecting_Perception_Failures_of_VQA_Models_Using_Metamorphic_CVPR_2021_paper.html
How Transferable Are Reasoning Patterns in VQA?https://openaccess.thecvf.com/content/CVPR2021/html/Kervadec_How_Transferable_Are_Reasoning_Patterns_in_VQA_CVPR_2021_paper.html
Domain-Robust VQA With Diverse Datasets and Methods but No Target Labelshttps://openaccess.thecvf.com/content/CVPR2021/html/Zhang_Domain-Robust_VQA_With_Diverse_Datasets_and_Methods_but_No_Target_CVPR_2021_paper.html
Learning Better Visual Dialog Agents With Pretrained Visual-Linguistic Representationhttps://openaccess.thecvf.com/content/CVPR2021/html/Tu_Learning_Better_Visual_Dialog_Agents_With_Pretrained_Visual-Linguistic_Representation_CVPR_2021_paper.html
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#iclr-2021
MultiModalQA: complex question answering over text, tables and imageshttps://openreview.net/pdf?id=ee6W5UgQLa
[page]https://allenai.github.io/multimodalqa/
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#naacl-hlt-2021
CLEVR_HYP: A Dataset and Baselines for Visual Question Answering with Hypothetical Actions over Imageshttps://arxiv.org/abs/2104.05981
[code]https://github.com/shailaja183/clevr_hyp
Video Question Answering with Phrases via Semantic Roleshttps://arxiv.org/abs/2104.03762
SOrT-ing VQA Models : Contrastive Gradient Learning for Improved Consistencyhttps://arxiv.org/abs/2010.10038
EaSe: A Diagnostic Tool for VQA based on Answer Diversityhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master
Ensemble of MRR and NDCG models for Visual Dialoghttps://arxiv.org/abs/2104.07511
[code]https://github.com/idansc/mrr-ndcg
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#aaai-2021
Regularizing Attention Networks for Anomaly Detection in Visual Question Answeringhttps://arxiv.org/abs/2009.10054
A Case Study of the Shortcut Effects in Visual Commonsense Reasoninghttps://www.aaai.org/AAAI21Papers/AAAI-9821.YeK.pdf
[code]https://github.com/yekeren/VCR-shortcut-effects-study
VisualMRC: Machine Reading Comprehension on Document Imageshttps://arxiv.org/abs/2101.11272
[page]https://github.com/nttmdlab-nlp/VisualMRC
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2020
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#emnlp-2020
MUTANT: A Training Paradigm for Out-of-Distribution Generalization in Visual Question Answeringhttps://www.aclweb.org/anthology/2020.emnlp-main.63/
[code]https://github.com/tejas-gokhale/vqa_mutant
Learning to Contrast the Counterfactual Samples for Robust Visual Question Answeringhttps://www.aclweb.org/anthology/2020.emnlp-main.265/
[code]https://github.com/jokieleung/CL-VQA
VD-BERT: A Unified Vision and Dialog Transformer with BERThttps://www.aclweb.org/anthology/2020.emnlp-main.269/
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#neurips-2020
Multimodal Graph Networks for Compositional Generalization in Visual Question Answeringhttps://papers.nips.cc/paper/2020/hash/1fd6c4e41e2c6a6b092eb13ee72bce95-Abstract.html
Removing Bias in Multi-modal Classifiers: Regularization by Maximizing Functional Entropieshttps://papers.nips.cc/paper/2020/hash/20d749bc05f47d2bd3026ce457dcfd8e-Abstract.html
Dialog without Dialog Data: Learning Visual Dialog Agents from VQA Datahttps://papers.nips.cc/paper/2020/hash/e7023ba77a45f7e84c5ee8a28dd63585-Abstract.html
On the Value of Out-of-Distribution Testing: An Example of Goodhart's Lawhttps://arxiv.org/abs/2005.09241
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#eccv-2020
Reducing Language Biases in Visual Question Answering with Visually-Grounded Question Encoderhttp://www.ecva.net/papers/eccv_2020/papers_ECCV/html/1765_ECCV_2020_paper.php
Knowledge-Based Video Question Answering with Unsupervised Scene Descriptionshttp://www.ecva.net/papers/eccv_2020/papers_ECCV/html/3056_ECCV_2020_paper.php
Semantic Equivalent Adversarial Data Augmentation for Visual Question Answeringhttp://www.ecva.net/papers/eccv_2020/papers_ECCV/html/3245_ECCV_2020_paper.php
Visual Question Answering on Image Setshttp://www.ecva.net/papers/eccv_2020/papers_ECCV/html/3640_ECCV_2020_paper.php
VQA-LOL: Visual Question Answering under the Lens of Logichttp://www.ecva.net/papers/eccv_2020/papers_ECCV/html/3742_ECCV_2020_paper.php
TRRNet: Tiered Relation Reasoning for Compositional Visual Question Answeringhttp://www.ecva.net/papers/eccv_2020/papers_ECCV/html/3752_ECCV_2020_paper.php
Spatially Aware Multimodal Transformers for TextVQAhttp://www.ecva.net/papers/eccv_2020/papers_ECCV/html/946_ECCV_2020_paper.php
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#cvpr-2020
Multi-Modal Graph Neural Network for Joint Reasoning on Vision and Scene Texthttp://arxiv.org/abs/2003.13962
[code]https://github.com/ricolike/mmgnn_textvqa
On the General Value of Evidence, and Bilingual Scene-Text Visual Question Answeringhttp://openaccess.thecvf.com/content_CVPR_2020/html/Wang_On_the_General_Value_of_Evidence_and_Bilingual_Scene-Text_Visual_CVPR_2020_paper.html
In Defense of Grid Features for Visual Question Answeringhttp://openaccess.thecvf.com/content_CVPR_2020/html/Jiang_In_Defense_of_Grid_Features_for_Visual_Question_Answering_CVPR_2020_paper.html
Counterfactual Samples Synthesizing for Robust Visual Question Answeringhttp://openaccess.thecvf.com/content_CVPR_2020/html/Chen_Counterfactual_Samples_Synthesizing_for_Robust_Visual_Question_Answering_CVPR_2020_paper.html
Counterfactual Vision and Language Learninghttp://openaccess.thecvf.com/content_CVPR_2020/html/Abbasnejad_Counterfactual_Vision_and_Language_Learning_CVPR_2020_paper.html
Iterative Answer Prediction With Pointer-Augmented Multimodal Transformers for TextVQAhttp://openaccess.thecvf.com/content_CVPR_2020/html/Hu_Iterative_Answer_Prediction_With_Pointer-Augmented_Multimodal_Transformers_for_TextVQA_CVPR_2020_paper.html
Towards Causal VQA: Revealing and Reducing Spurious Correlations by Invariant and Covariant Semantic Editinghttp://openaccess.thecvf.com/content_CVPR_2020/html/Agarwal_Towards_Causal_VQA_Revealing_and_Reducing_Spurious_Correlations_by_Invariant_CVPR_2020_paper.html
SQuINTing at VQA Models: Introspecting VQA Models With Sub-Questionshttp://openaccess.thecvf.com/content_CVPR_2020/html/Selvaraju_SQuINTing_at_VQA_Models_Introspecting_VQA_Models_With_Sub-Questions_CVPR_2020_paper.html
TA-Student VQA: Multi-Agents Training by Self-Questioninghttp://openaccess.thecvf.com/content_CVPR_2020/html/Xiong_TA-Student_VQA_Multi-Agents_Training_by_Self-Questioning_CVPR_2020_paper.html
VQA With No Questions-Answers Traininghttp://openaccess.thecvf.com/content_CVPR_2020/html/Vatashsky_VQA_With_No_Questions-Answers_Training_CVPR_2020_paper.html
Hierarchical Conditional Relation Networks for Video Question Answeringhttp://openaccess.thecvf.com/content_CVPR_2020/html/Le_Hierarchical_Conditional_Relation_Networks_for_Video_Question_Answering_CVPR_2020_paper.html
Modality Shifting Attention Network for Multi-Modal Video Question Answeringhttp://openaccess.thecvf.com/content_CVPR_2020/html/Kim_Modality_Shifting_Attention_Network_for_Multi-Modal_Video_Question_Answering_CVPR_2020_paper.html
Webly Supervised Knowledge Embedding Model for Visual Reasoninghttp://openaccess.thecvf.com/content_CVPR_2020/html/Zheng_Webly_Supervised_Knowledge_Embedding_Model_for_Visual_Reasoning_CVPR_2020_paper.html
Differentiable Adaptive Computation Time for Visual Reasoninghttp://openaccess.thecvf.com/content_CVPR_2020/html/Eyzaguirre_Differentiable_Adaptive_Computation_Time_for_Visual_Reasoning_CVPR_2020_paper.html
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#acl-2020
A negative case analysis of visual grounding methods for VQAhttps://arxiv.org/abs/2004.05704
Cross-Modality Relevance for Reasoning on Language and Visionhttps://arxiv.org/abs/2005.06035
Dense-Caption Matching and Frame-Selection Gating for Temporal Localization in VideoQAhttps://arxiv.org/abs/2005.06409
TVQA+: Spatio-Temporal Grounding for Video Question Answeringhttps://arxiv.org/abs/1904.11574
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#wacv-2020
BERT representations for Video Question Answeringhttp://openaccess.thecvf.com/content_WACV_2020/papers/Yang_BERT_representations_for_Video_Question_Answering_WACV_2020_paper.pdf
Deep Bayesian Network for Visual Question Generationhttp://openaccess.thecvf.com/content_WACV_2020/html/Patro_Deep_Bayesian_Network_for_Visual_Question_Generation_WACV_2020_paper.html
Robust Explanations for Visual Question Answeringhttp://openaccess.thecvf.com/content_WACV_2020/html/Patro_Robust_Explanations_for_Visual_Question_Answering_WACV_2020_paper.html
Visual Question Answering on 360deg Imageshttp://openaccess.thecvf.com/content_WACV_2020/html/Chou_Visual_Question_Answering_on_360deg_Images_WACV_2020_paper.html
LEAF-QA: Locate, Encode & Attend for Figure Question Answeringhttp://openaccess.thecvf.com/content_WACV_2020/html/Chaudhry_LEAF-QA_Locate_Encode__Attend_for_Figure_Question_Answering_WACV_2020_paper.html
Answering Questions about Data Visualizations using Efficient Bimodal Fusionhttp://openaccess.thecvf.com/content_WACV_2020/html/Kafle_Answering_Questions_about_Data_Visualizations_using_Efficient_Bimodal_Fusion_WACV_2020_paper.html
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#aaai-2020
Multi‐Question Learning for Visual Question Answeringhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master
Explanation vs Attention: A Two-Player Game to Obtain Attention for VQAhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master
Overcoming Language Priors in VQA via Decomposed Linguistic Representationshttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master
Unified Vision-Language Pre-Training for Image Captioning and VQAhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master
Re‐Attention for Visual Question Answeringhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master
Divide and Conquer: Question­‐Guided Spatio­‐Temporal Contextual Attention for Video Question Answeringhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master
Reasoning with Heterogeneous Graph Alignment for Video Question Answeringhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master
Location­‐aware Graph Convolutional Networks for Video Question Answeringhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master
KnowIT VQA: Answering Knowledge­‐Based Questions about Videoshttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/blob/master
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2019
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#acl-2019
Generating Question Relevant Captions to Aid Visual Question Answeringhttps://www.aclweb.org/anthology/P19-1348.pdf
Psycholinguistics Meets Continual Learning: Measuring Catastrophic Forgetting in Visual Question Answeringhttps://www.aclweb.org/anthology/P19-1350.pdf
[code]https://www.aclweb.org/anthology/P19-1350/
Multi-grained Attention with Object-level Grounding for Visual Question Answeringhttps://www.aclweb.org/anthology/P19-1349.pdf
Improving Visual Question Answering by Referring to Generated Paragraph Captionshttps://www.aclweb.org/anthology/P19-1351.pdf
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#iccv-2019
Compact Trilinear Interaction for Visual Question Answeringhttp://openaccess.thecvf.com/content_ICCV_2019/html/Do_Compact_Trilinear_Interaction_for_Visual_Question_Answering_ICCV_2019_paper.html
Scene Text Visual Question Answeringhttp://openaccess.thecvf.com/content_ICCV_2019/html/Biten_Scene_Text_Visual_Question_Answering_ICCV_2019_paper.html
Multi-Modality Latent Interaction Network for Visual Question Answeringhttp://openaccess.thecvf.com/content_ICCV_2019/html/Gao_Multi-Modality_Latent_Interaction_Network_for_Visual_Question_Answering_ICCV_2019_paper.html
Relation-Aware Graph Attention Network for Visual Question Answeringhttp://openaccess.thecvf.com/content_ICCV_2019/html/Li_Relation-Aware_Graph_Attention_Network_for_Visual_Question_Answering_ICCV_2019_paper.html
Why Does a Visual Question Have Different Answers?http://openaccess.thecvf.com/content_ICCV_2019/papers/Bhattacharya_Why_Does_a_Visual_Question_Have_Different_Answers_ICCV_2019_paper.pdf
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#neurips-2019
RUBi: Reducing Unimodal Biases for Visual Question Answeringhttps://papers.nips.cc/paper/8371-rubi-reducing-unimodal-biases-for-visual-question-answering
Self-Critical Reasoning for Robust Visual Question Answeringhttps://papers.nips.cc/paper/9066-self-critical-reasoning-for-robust-visual-question-answering
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#cvpr-2019
Deep Modular Co-Attention Networks for Visual Question Answeringhttp://openaccess.thecvf.com/content_CVPR_2019/papers/Yu_Deep_Modular_Co-Attention_Networks_for_Visual_Question_Answering_CVPR_2019_paper.pdf
[code]https://github.com/MILVLG/mcan-vqa
Information Maximizing Visual Question Generationhttps://arxiv.org/abs/1903.11207
Social-IQ: A Question Answering Benchmark for Artificial Social Intelligencehttp://openaccess.thecvf.com/content_CVPR_2019/papers/Zadeh_Social-IQ_A_Question_Answering_Benchmark_for_Artificial_Social_Intelligence_CVPR_2019_paper.pdf
Learning to Compose Dynamic Tree Structures for Visual Contextshttps://arxiv.org/abs/1812.01880
Transfer Learning via Unsupervised Task Discovery for Visual Question Answeringhttps://arxiv.org/abs/1810.02358
Video Relationship Reasoning using Gated Spatio-Temporal Energy Graphhttps://arxiv.org/abs/1903.10547
[code]https://github.com/yaohungt/Gated-Spatio-Temporal-Energy-Graph
Explainable and Explicit Visual Reasoning over Scene Graphshttps://arxiv.org/abs/1812.01855
[code]https://github.com/shijx12/XNM-Net
MUREL: Multimodal Relational Reasoning for Visual Question Answeringhttps://arxiv.org/abs/1902.09487
[code]https://github.com/Cadene/murel.bootstrap.pytorch
Image-Question-Answer Synergistic Network for Visual Dialoghttps://arxiv.org/abs/1902.09774
RAVEN: A Dataset for Relational and Analogical Visual rEasoNinghttps://arxiv.org/abs/1903.02741
[project page]http://wellyzhang.github.io/project/raven.html
Cycle-Consistency for Robust Visual Question Answeringhttp://openaccess.thecvf.com/content_CVPR_2019/html/Shah_Cycle-Consistency_for_Robust_Visual_Question_Answering_CVPR_2019_paper.html
It's Not About the Journey; It's About the Destination: Following Soft Paths Under Question-Guidance for Visual Reasoninghttp://openaccess.thecvf.com/content_CVPR_2019/html/Haurilet_Its_Not_About_the_Journey_Its_About_the_Destination_Following_CVPR_2019_paper.html
OK-VQA: A Visual Question Answering Benchmark Requiring External Knowledgehttp://openaccess.thecvf.com/content_CVPR_2019/html/Marino_OK-VQA_A_Visual_Question_Answering_Benchmark_Requiring_External_Knowledge_CVPR_2019_paper.html
Visual Question Answering as Reading Comprehensionhttp://openaccess.thecvf.com/content_CVPR_2019/html/Li_Visual_Question_Answering_as_Reading_Comprehension_CVPR_2019_paper.html
Dynamic Fusion With Intra- and Inter-Modality Attention Flow for Visual Question Answeringhttp://openaccess.thecvf.com/content_CVPR_2019/html/Gao_Dynamic_Fusion_With_Intra-_and_Inter-Modality_Attention_Flow_for_Visual_CVPR_2019_paper.html
Explicit Bias Discovery in Visual Question Answering Modelshttp://openaccess.thecvf.com/content_CVPR_2019/html/Manjunatha_Explicit_Bias_Discovery_in_Visual_Question_Answering_Models_CVPR_2019_paper.html
Answer Them All! Toward Universal Visual Question Answering Modelshttp://openaccess.thecvf.com/content_CVPR_2019/html/Shrestha_Answer_Them_All_Toward_Universal_Visual_Question_Answering_Models_CVPR_2019_paper.html
Visual Query Answering by Entity-Attribute Graph Matching and Reasoninghttp://openaccess.thecvf.com/content_CVPR_2019/html/Xiong_Visual_Query_Answering_by_Entity-Attribute_Graph_Matching_and_Reasoning_CVPR_2019_paper.html
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#aaai-2019
Differential Networks for Visual Question Answeringhttps://www.aaai.org/Papers/AAAI/2019/AAAI-WuC.76.pdf
BLOCK: Bilinear Superdiagonal Fusion for Visual Question Answering and Visual Relationship Detectionhttps://arxiv.org/abs/1902.00038
[code]https://github.com/Cadene/block.bootstrap.pytorch
Dynamic Capsule Attention for Visual Question Answeringhttps://www.aaai.org/Papers/AAAI/2019/AAAI-ZhouYiyi2.3610.pdf
[code]https://github.com/XMUVQA/CapsAtt
Beyond RNNs: Positional Self-Attention with Co-Attention for Video Question Answeringhttps://www.semanticscholar.org/paper/Beyond-RNNs%3A-Positional-Self-Attention-with-for-Li-Song/565359aac8914505e6b02db05822ee63d3ffd03a
[code]https://github.com/lixiangpengcs/PSAC
[code]https://github.com/sanket0211/WK-VQA
Free VQA Models from Knowledge Inertia by Pairwise Inconformity Learninghttps://www.aaai.org/Papers/AAAI/2019/AAAI-ZhouYiyi1.1233.pdf
[code]https://github.com/xiangmingLi/PIL
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#other
Focal Visual-Text Attention for Memex Question Answeringhttps://ieeexplore.ieee.org/abstract/document/8603827/
[code]https://memexqa.cs.cmu.edu/
Combining Multiple Cues for Visual Madlibs Question Answeringhttps://arxiv.org/abs/1611.00393
Large-Scale Answerer in Questioner's Mind for Visual Dialog Question Generationhttps://arxiv.org/abs/1902.08355
[code]https://github.com/naver/aqm-plus
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2018
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#nips-2018
Bilinear Attention Networkshttps://papers.nips.cc/paper/7429-bilinear-attention-networks
Chain of Reasoning for Visual Question Answeringhttps://papers.nips.cc/paper/7311-chain-of-reasoning-for-visual-question-answering
Learning Conditioned Graph Structures for Interpretable Visual Question Answeringhttps://papers.nips.cc/paper/8054-learning-conditioned-graph-structures-for-interpretable-visual-question-answering
[code]https://github.com/aimbrain/vqa-project
Learning to Specialize with Knowledge Distillation for Visual Question Answeringhttps://papers.nips.cc/paper/8031-learning-to-specialize-with-knowledge-distillation-for-visual-question-answering
Out of the Box: Reasoning with Graph Convolution Nets for Factual Visual Question Answeringhttps://papers.nips.cc/paper/7531-out-of-the-box-reasoning-with-graph-convolution-nets-for-factual-visual-question-answering
Overcoming Language Priors in Visual Question Answering with Adversarial Regularizationhttps://papers.nips.cc/paper/7427-overcoming-language-priors-in-visual-question-answering-with-adversarial-regularization
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#aaai-2018
Explicit Reasoning over End-to-End Neural Architectures for Visual Question Answeringhttps://aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/16446
Co-Attending Free-Form Regions and Detections with Multi-Modal Multiplicative Feature Embedding for Visual Question Answeringhttps://aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/16249
[code]https://github.com/lupantech/dual-mfa-vqa
Exploring Human-Like Attention Supervision in Visual Question Answeringhttps://aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/16485
Movie Question Answering: Remembering the Textual Cues for Layered Visual Contentshttps://aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/16359
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#ijcai-2018
Feature Enhancement in Attention for Visual Question Answeringhttps://www.ijcai.org/proceedings/2018/586
A Question Type Driven Framework to Diversify Visual Question Generationhttps://www.ijcai.org/proceedings/2018/563
Multi-Turn Video Question Answering via Multi-Stream Hierarchical Attention Context Networkhttps://www.ijcai.org/proceedings/2018/513
Open-Ended Long-form Video Question Answering via Adaptive Hierarchical Reinforced Networkshttps://www.ijcai.org/proceedings/2018/512
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#cvpr-2018
Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Anderson_Bottom-Up_and_Top-Down_CVPR_2018_paper.html
[code(author)]https://github.com/peteanderson80/bottom-up-attention
[code(pythiaV0.1)]https://github.com/facebookresearch/pythia
[code(Pytorch Reimplementation)]https://github.com/hengyuan-hu/bottom-up-attention-vqa
Tips and Tricks for Visual Question Answering: Learnings From the 2017 Challengehttp://openaccess.thecvf.com/content_cvpr_2018/html/Teney_Tips_and_Tricks_CVPR_2018_paper.html
Learning by Asking Questionshttp://openaccess.thecvf.com/content_cvpr_2018/html/Misra_Learning_by_Asking_CVPR_2018_paper.html
Embodied Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Das_Embodied_Question_Answering_CVPR_2018_paper.html
VizWiz Grand Challenge: Answering Visual Questions From Blind Peoplehttp://openaccess.thecvf.com/content_cvpr_2018/html/Gurari_VizWiz_Grand_Challenge_CVPR_2018_paper.html
Textbook Question Answering Under Instructor Guidance With Memory Networkshttp://openaccess.thecvf.com/content_cvpr_2018/html/Li_Textbook_Question_Answering_CVPR_2018_paper.html
[code]https://github.com/freerailway/igmn
IQA: Visual Question Answering in Interactive Environmentshttp://openaccess.thecvf.com/content_cvpr_2018/html/Gordon_IQA_Visual_Question_CVPR_2018_paper.html
[sample video]https://youtu.be/pXd3C-1jr98
Don't Just Assume; Look and Answer: Overcoming Priors for Visual Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Agrawal_Dont_Just_Assume_CVPR_2018_paper.html
Learning Answer Embeddings for Visual Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Hu_Learning_Answer_Embeddings_CVPR_2018_paper.html
DVQA: Understanding Data Visualizations via Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Kafle_DVQA_Understanding_Data_CVPR_2018_paper.html
Cross-Dataset Adaptation for Visual Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Chao_Cross-Dataset_Adaptation_for_CVPR_2018_paper.html
Two Can Play This Game: Visual Dialog With Discriminative Question Generation and Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Jain_Two_Can_Play_CVPR_2018_paper.html
Improved Fusion of Visual and Language Representations by Dense Symmetric Co-Attention for Visual Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Nguyen_Improved_Fusion_of_CVPR_2018_paper.html
Visual Question Generation as Dual Task of Visual Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Li_Visual_Question_Generation_CVPR_2018_paper.html
Focal Visual-Text Attention for Visual Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Liang_Focal_Visual-Text_Attention_CVPR_2018_paper.html
Motion-Appearance Co-Memory Networks for Video Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Gao_Motion-Appearance_Co-Memory_Networks_CVPR_2018_paper.html
Visual Question Answering With Memory-Augmented Networkshttp://openaccess.thecvf.com/content_cvpr_2018/html/Ma_Visual_Question_Answering_CVPR_2018_paper.html
Visual Question Reasoning on General Dependency Treehttp://openaccess.thecvf.com/content_cvpr_2018/html/Cao_Visual_Question_Reasoning_CVPR_2018_paper.html
Differential Attention for Visual Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Patro_Differential_Attention_for_CVPR_2018_paper.html
Learning Visual Knowledge Memory Networks for Visual Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Su_Learning_Visual_Knowledge_CVPR_2018_paper.html
IVQA: Inverse Visual Question Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Liu_IVQA_Inverse_Visual_CVPR_2018_paper.html
Customized Image Narrative Generation via Interactive Visual Question Generation and Answeringhttp://openaccess.thecvf.com/content_cvpr_2018/html/Shin_Customized_Image_Narrative_CVPR_2018_paper.html
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#acm-mm-2018
Object-Difference Attention: A simple relational attention for Visual Question Answeringhttps://dl.acm.org/citation.cfm?doid=3240508.3240513
Enhancing Visual Question Answering Using Dropouthttps://doi.org/10.1145/3240508.3240662
Fast Parameter Adaptation for Few-shot Image Captioning and Visual Question Answeringhttps://doi.org/10.1145/3240508.3240527
[code]https://github.com/D-X-Y/FPAIT
Explore Multi-Step Reasoning in Video Question Answeringhttps://doi.org/10.1145/3240508.3240563
[code]https://github.com/SVQA-founder/SVQA/tree/master/code
[SVQA dataset]https://svqa-founder.github.io/SVQA/
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#eccv-2018
Visual Question Answering as a Meta Learning Taskhttp://openaccess.thecvf.com/content_ECCV_2018/html/Damien_Teney_Visual_Question_Answering_ECCV_2018_paper.html
Question-Guided Hybrid Convolution for Visual Question Answeringhttp://openaccess.thecvf.com/content_ECCV_2018/html/gao_peng_Question-Guided_Hybrid_Convolution_ECCV_2018_paper.html
Goal-Oriented Visual Question Generation via Intermediate Rewardshttp://openaccess.thecvf.com/content_ECCV_2018/html/Junjie_Zhang_Goal-Oriented_Visual_Question_ECCV_2018_paper.html
Multimodal Dual Attention Memory for Video Story Question Answeringhttp://openaccess.thecvf.com/content_ECCV_2018/html/Kyungmin_Kim_Multimodal_Dual_Attention_ECCV_2018_paper.html
A Joint Sequence Fusion Model for Video Question Answering and Retrievalhttp://openaccess.thecvf.com/content_ECCV_2018/html/Youngjae_Yu_A_Joint_Sequence_ECCV_2018_paper.html
Deep Attention Neural Tensor Network for Visual Question Answeringhttp://openaccess.thecvf.com/content_ECCV_2018/html/Yalong_Bai_Deep_Attention_Neural_ECCV_2018_paper.html
Question Type Guided Attention in Visual Question Answeringhttp://openaccess.thecvf.com/content_ECCV_2018/html/Yang_Shi_Question_Type_Guided_ECCV_2018_paper.html
Learning Visual Question Answering by Bootstrapping Hard Attentionhttp://openaccess.thecvf.com/content_ECCV_2018/html/Mateusz_Malinowski_Learning_Visual_Question_ECCV_2018_paper.html
Straight to the Facts: Learning Knowledge Base Retrieval for Factual Visual Question Answeringhttp://openaccess.thecvf.com/content_ECCV_2018/html/Medhini_Gulganjalli_Narasimhan_Straight_to_the_ECCV_2018_paper.html
Visual Question Generation for Class Acquisition of Unknown Objectshttp://openaccess.thecvf.com/content_ECCV_2018/html/Kohei_Uehara_Visual_Question_Generation_ECCV_2018_paper.html
[code]https://github.com/mil-tokyo/vqg-unknown
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#other-1
Image Captioning and Visual Question Answering Based on Attributes and External Knowledgehttps://arxiv.org/abs/1603.02814
FVQA: Fact-Based Visual Question Answeringhttps://arxiv.org/abs/1606.05433
R-VQA: Learning Visual Relation Facts with Semantic Attention for Visual Question Answeringhttps://dl.acm.org/citation.cfm?doid=3219819.3220036
[code(Dataset)]https://github.com/lupantech/rvqa
Interpretable Counting for Visual Question Answeringhttps://arxiv.org/abs/1712.08697
Learning to Count Objects in Natural Images for Visual Question Answeringhttps://openreview.net/forum?id=B12Js_yRb
A Better Way to Attend: Attention With Trees for Video Question Answeringhttps://ieeexplore.ieee.org/document/8419716
[code]https://github.com/xuehy/TreeAttention
Zero-Shot Transfer VQA Datasethttps://arxiv.org/abs/1811.00692
Visual Question Answering using Explicit Visual Attentionhttps://ieeexplore.ieee.org/abstract/document/8351158/
Explicit ensemble attention learning for improving visual question answeringhttps://www.sciencedirect.com/science/article/abs/pii/S0167865518301600
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#2017-2015
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#other-2
awesome-vqahttps://github.com/JamesChuanggg/awesome-vqa
JamesChuanggghttps://github.com/JamesChuanggg
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#iccv-2017
Learning to Reason: End-to-End Module Networks for Visual Question Answeringhttps://ieeexplore.ieee.org/document/8237355
Structured Attentions for Visual Question Answeringhttps://ieeexplore.ieee.org/document/8237407
[code]https://github.com/shtechair/vqa-sva
VQS: Linking Segmentations to Questions and Answers for Supervised Attention in VQA and Question-Focused Semantic Segmentationhttps://ieeexplore.ieee.org/document/8237463
[code]https://github.com/Cold-Winter/vqs
Multi-modal Factorized Bilinear Pooling with Co-attention Learning for Visual Question Answeringhttps://ieeexplore.ieee.org/document/8237464
[code]https://github.com/yuzcccc/vqa-mfb
An Analysis of Visual Question Answering Algorithmshttps://ieeexplore.ieee.org/document/8237479
MUTAN: Multimodal Tucker Fusion for Visual Question Answeringhttps://ieeexplore.ieee.org/document/8237547
[code]https://github.com/cadene/vqa.pytorch
MarioQA: Answering Questions by Watching Gameplay Videoshttps://ieeexplore.ieee.org/document/8237574
Learning to Disambiguate by Asking Discriminative Questionshttps://ieeexplore.ieee.org/document/8237632
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#vqa-challenge-leaderboard
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#test-std-2018
VQA Challenge 2018 Leaderboard in EvalAIhttps://evalai.cloudcv.org/featured-challenges/80/leaderboard/124
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#test-std-2017
VQA Challenge 2017(Open-Ended) Leaderboard in EvalAIhttps://evalai.cloudcv.org/featured-challenges/1/leaderboard/3
TextVQAhttps://textvqa.org/
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#textvqa
TextVQA Challenge 2019 Leaderboard in EvalAIhttps://evalai.cloudcv.org/web/challenges/challenge-page/244/overview
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#vqa-cp
The VQA-CP leaderboardhttps://github.com/cdancette/vqa-cp-leaderboard
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#licenses
http://creativecommons.org/publicdomain/zero/1.0/
Jokie Leunghttps://github.com/jokieleung
https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#reference-and-acknowledgement
awesome-image-captioninghttps://github.com/zhjohnchan/awesome-image-captioning
Zhihong Chenhttps://github.com/zhjohnchan
awesome-vqahttps://github.com/JamesChuanggg/awesome-vqa
JamesChuanggghttps://github.com/JamesChuanggg
Readme https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#readme-ov-file
Contributing https://patch-diff.githubusercontent.com/billiecn/visual-question-answering#contributing-ov-file
Please reload this pagehttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering
Activityhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/activity
1 starhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/stargazers
0 watchinghttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/watchers
1 forkhttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/forks
Report repository https://patch-diff.githubusercontent.com/contact/report-content?content_url=https%3A%2F%2Fgithub.com%2Fbilliecn%2Fvisual-question-answering&report=billiecn+%28user%29
Releaseshttps://patch-diff.githubusercontent.com/billiecn/visual-question-answering/releases
Packages 0https://patch-diff.githubusercontent.com/users/billiecn/packages?repo_name=visual-question-answering
https://github.com
Termshttps://docs.github.com/site-policy/github-terms/github-terms-of-service
Privacyhttps://docs.github.com/site-policy/privacy-policies/github-privacy-statement
Securityhttps://github.com/security
Statushttps://www.githubstatus.com/
Communityhttps://github.community/
Docshttps://docs.github.com/
Contacthttps://support.github.com?tags=dotcom-footer

Viewport: width=device-width


URLs of crawlers that visited me.