Skip to content
Permalink

Comparing changes

Choose two branches to see what’s changed or to start a new pull request. If you need to, you can also or learn more about diff comparisons.

Open a pull request

Create a new pull request by comparing changes across two branches. If you need to, you can also . Learn more about diff comparisons here.
base repository: GT-Vision-Lab/VQA_LSTM_CNN
Failed to load repositories. Confirm that selected base ref is valid, then try again.
Loading
base: v1.0
Choose a base ref
...
head repository: GT-Vision-Lab/VQA_LSTM_CNN
Failed to load repositories. Confirm that selected head ref is valid, then try again.
Loading
compare: master
Choose a head ref
  • 4 commits
  • 1 file changed
  • 1 contributor

Commits on Jun 2, 2016

  1. Update readme.md

    jiasenlu committed Jun 2, 2016
    Copy the full SHA
    7b2ede3 View commit details

Commits on Jun 3, 2016

  1. Update readme.md

    jiasenlu committed Jun 3, 2016
    Copy the full SHA
    31ded04 View commit details

Commits on Jun 7, 2016

  1. Update readme.md

    jiasenlu committed Jun 7, 2016
    Copy the full SHA
    b32beeb View commit details

Commits on Jun 14, 2016

  1. Update readme.md

    jiasenlu authored Jun 14, 2016
    Copy the full SHA
    15d7d4b View commit details
Showing with 3 additions and 2 deletions.
  1. +3 −2 readme.md
5 changes: 3 additions & 2 deletions readme.md
Original file line number Diff line number Diff line change
@@ -2,6 +2,7 @@

Train a deeper LSTM and normalized CNN Visual Question Answering model. This current code can get **58.16** on Open-Ended and **63.09** on Multiple-Choice on **test-standard** split. You can check [Codalab leaderboard](https://competitions.codalab.org/competitions/6961#results) for more details.

**New VQA Model with better performance and cleaner code can be found here [https://github.com/jiasenlu/HieCoAttenVQA](https://github.com/jiasenlu/HieCoAttenVQA)**

### Requirements

@@ -82,7 +83,7 @@ with the default parameter, this will take several hours on a sinlge Tesla k40 G
If you use this code as part of any published research, please acknowledge the following repo
```
@misc{Lu2015,
author = {Jiasen Lu, Xiao Lin, Dhruv Batra and Devi Parikh},
author = {Jiasen Lu and Xiao Lin and Dhruv Batra and Devi Parikh},
title = {Deeper LSTM and normalized CNN Visual Question Answering model},
year = {2015},
publisher = {GitHub},
@@ -94,7 +95,7 @@ commit = {6c91cb9}
If you use the VQA dataset as part of any published research, please acknowledge the following paper
```
@InProceedings{Antol_2015_ICCV,
author = {Antol, Stanislaw and Agrawal, Aishwarya and Lu, Jiasen and Mitchell, Margaret and Batra, Dhruv and Lawrence Zitnick, C. and Parikh, Devi},
author = {Stanislaw Antol and Aishwarya Agrawal and Jiasen Lu and Margaret Mitchell and Dhruv Batra and C. Lawrence Zitnick and Devi Parikh},
title = {VQA: Visual Question Answering},
booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
month = {December},