Skip to content

Latest commit

 

History

History
30 lines (22 loc) · 1.38 KB

File metadata and controls

30 lines (22 loc) · 1.38 KB

NLP benchmark

As it was found out WBF method in one-dimensional variant works good for NER tasks in NLP problems. Here you can find benchmark which is based on Feedback Prize - Evaluating Student Writing Kaggle competition. Benchmark consists of OOF predictions of 10 different NLP models for the competition dataset. WBF method allowed achieve 2nd place in competition. Credits for benchmarks and idea of using WBF for NLP task goes to Chris Deotte and Udbhav Bamba.

Model Validation score
lsg-large 0.7026
longformer-lstm 0.7024
deberta-jaccard 0.6990
deberta-large-v3 0.6945
deberta-xlarge-v2 0.6955
bird-base-1024 0.6746
deberta-large 0.6950
deberta-xlarge 0.6984
funnel-large 0.6880
yoso 0.6516

Benchmark files

Download ~16 MB

Ensemble results

There is python code to get high score on validation using WBF method: run_benchmark_nlp.py

WBF with IoU = 0.33 gives 0.7403 on validation (best model before ensemble gives only 0.7026).