mcQA - Multiple Choice Question Answering
mcQA is a multiple choice question answering python library, using Language Models.
Installation
With pip
pip install mcqa
From source
git clone https://github.com/mcqa-suite/mcqa.git
cd mcQA
pip install -e .
Getting started
Data preparation
To train a mcQA
model, you need to create a csv file with n+2 columns, n being the number of choices for each question. The first column should be the context sentence, the n following columns should be the choices for that question and the last column is the selected answer.
Below is an example of a 3 choice question (taken from the CoS-E dataset) :
Context sentence | Choice 1 | Choice 2 | Choice 3 | Label |
---|---|---|---|---|
People do what during their time off from work? | take trips | brow shorter | become hysterical | take trips |
If you have a trained mcQA
model and want to infer on a dataset, it should have the same format as the train data, but the label
column.
See example data preparation below:
from mcqa.data import MCQAData
mcqa_data = MCQAData(bert_model="bert-base-uncased", lower_case=True, max_seq_length=256)
train_dataset = mcqa_data.read(data_file='swagaf/data/train.csv', is_training=True)
test_dataset = mcqa_data.read(data_file='swagaf/data/test.csv', is_training=False)
Model training
from mcqa.models import Model
mdl = Model(bert_model="bert-base-uncased", device="cuda")
mdl.fit(train_dataset, train_batch_size=32, num_train_epochs=20)
Prediction
preds = mdl.predict(test_dataset, eval_batch_size=32)
Evaluation
from sklearn.metrics import accuracy_score
from mcqa.data import get_labels
print(accuracy_score(preds, get_labels(train_dataset)))
References
- Explain Yourself! Leveraging Language Models for Commonsense Reasoning
- SWAG: A Large-Scale Adversarial Dataset for Grounded Commonsense Inference
- Are We Modeling the Task or the Annotator? An Investigation of Annotator Bias in Natural Language Understanding Datasets
- Can a Suit of Armor Conduct Electricity? A New Dataset for Open Book Question Answering
- CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge
- RACE: Large-scale ReAding Comprehension Dataset From Examinations
- Stanford CS224N: NLP with Deep Learning Lecture 10 – Question Answering