CSpell

Performance Tests on Other Test Sets

I. MEDLINE Title

  • Informally tested to correct spelling errors in MEDLINE titles
  • Test file includes 2585 MEDLINE titles with errors
  • The preliminary results are similar to those on the consumer health questions
  • Indicating that CSpell might potentially be used for other biomedical spelling error corrections.
  • Initial result for correction (10 instances):
    PrecisionRecallF1
    1.000.600.75

II. CSpell Test Set on the important terms

  • Data: Test Set
  • Evaluated its performance on only those terms that are important for question understanding.
  • The important terms were identified during manual annotation.
  • The results on the important terms are very close to the overall results.

    TestsPrecisionRecallF1
    Non-word Detection0.89370.82590.8585
    Non-word Correction0.73430.67860.7053
    Real-word Detection0.91790.74220.8207
    Real-word Correction0.75360.60940.6739