Here is the utility of the various files:
-
demo_batch.py
: You need access to pretrained models (included in the repo to run this example) -
get_started.sh
: Downloads data, VQAtools, pre-computed features, and trains a model. Run this script when you are done with the dependencies. -
dumpText.py
: Dumps the questions and answers from the VQA json files to some text files for later ease of use. Runpython dumpText.py -h
for more info. -
trainMLP.py
: Trains Multi-Layer perceptrons. Runpython trainMLP.py -h
for more info. -
trainLSTM_1.py
: Trains LSTM-based model. Runpython trainLSTM_1.py -h
for more info. -
trainLSTM_language.py
: Trains LSTM-based language-only model. Runpython trainLSTM_language.py -h
for more info. -
evaluateMLP.py
: Evaluates models trained bytrainMLP.py
. Needs model json file, hdf5 weights file, and output txt file destinations to run. -
evaluateLSTM.py
: Evaluates models trained bytrainLSTM_1.py
andtrainLSTM_language.py
. Needs model json file, hdf5 weights file, and output txt file destinations to run. -
features.py
: Contains functions that are used to convert images and words to vectors (or sequences of vectors). -
utils.py
: Exactly what you think. -
own_image.py
: Use your own image. Caffe installation required -
extract_features.py
: Extract 4096D VGG features from a VGG Caffe Model -
vgg_features.prototxt
: VGG Caffe Model Definition