Reading Comprehension

Focus on Service orientied metrics (eg. 1-example inference latency)

  • Exists samples in reports/summary/ directory

SQuAD v1.1

Model Inference Latency
(1-example/ms)
F1 (SQuAD) BaseConfig Note
BiDAF 142.644 cpu / 32.545 gpu 77.747 squad/bidaf.json -
BiDAF + ELMo - cpu / - gpu 82.288 squad/bidaf+elmo.json -
DrQA - cpu / - gpu 77.049 squad/drqa.json -
DocQA - cpu / - gpu 80.635 squad/docqa.json -
DocQA + ELMo - cpu / - gpu 84.372 squad/docqa+elmo.json -
QANet - cpu / - gpu 79.800 squad/qanet.json -
BERT - cpu / - gpu 87.130 squad/bert_base-_uncased.json -

Plot

  • Inference Latency (1-example)

images