Reading Comprehension¶
Focus on Service orientied metrics (eg. 1-example inference latency)
Exists samples in
reports/summary/
directory
SQuAD v1.1¶
Model | Inference Latency (1-example/ms) |
F1 (SQuAD) | BaseConfig | Note |
---|---|---|---|---|
BiDAF | 142.644 cpu / 32.545 gpu |
77.747 | squad/bidaf.json | - |
BiDAF + ELMo | - cpu / - gpu |
82.288 | squad/bidaf+elmo.json | - |
DrQA | - cpu / - gpu |
77.049 | squad/drqa.json | - |
DocQA | - cpu / - gpu |
80.635 | squad/docqa.json | - |
DocQA + ELMo | - cpu / - gpu |
84.372 | squad/docqa+elmo.json | - |
QANet | - cpu / - gpu |
79.800 | squad/qanet.json | - |
BERT | - cpu / - gpu |
87.130 | squad/bert_base-_uncased.json | - |
Plot¶
Inference Latency (1-example)