MUC-7 (State of the art): Difference between revisions

From ACL Wiki
Jump to navigation Jump to search
Pythonner (talk | contribs)
No edit summary
Pythonner (talk | contribs)
No edit summary
Line 15: Line 15:
! System name
! System name
! Short description
! Short description
! System type
! Main publications
! Main publications
! Software
! Software
! Results (F)
! Results (F)
|-
|-
| Human
| Annotator
| Human annotator
| Human annotator
| -
| [http://www.itl.nist.gov/iad/894.02/related_projects/muc/proceedings/muc_7_toc.html MUC-7 proceedings]
| [http://www.itl.nist.gov/iad/894.02/related_projects/muc/proceedings/muc_7_toc.html MUC-7 proceedings]
|  
| -
| 97.60%
| 97.60%
|-
|-
| LTG
| LTG
| Best MUC-7 participant
| Best MUC-7 participant
| H
| Mikheev, Grover and Moens (1998)
| Mikheev, Grover and Moens (1998)
|  
| -
| 93.39%
| 93.39%
|-
| Baseline
| Vocabulary transfer from training to testing
| S
| Whitelaw and Patrick (2003)
| -
| 58.89%
|-
|-
|}
|}
* '''System type''': R = hand-crafted rules, S = supervised learning, U = unsupervised learning, H = hybrid




== References ==
== References ==


Mikheev, A., Grover, C., and Moens, M. (1998). [http://www-nlpir.nist.gov/related_projects/muc/proceedings/muc_7_proceedings/ltg_muc7.pdf Description of the LTG system used for MUC-7]. ''Proceedings of the Seventh Message Understanding Conference (MUC-7)''. Fairfax, Virginia.
Mikheev, A., Grover, C. and Moens, M. (1998). [http://www-nlpir.nist.gov/related_projects/muc/proceedings/muc_7_proceedings/ltg_muc7.pdf Description of the LTG system used for MUC-7]. ''Proceedings of the Seventh Message Understanding Conference (MUC-7)''. Fairfax, Virginia.
 
Whitelaw, C. and Patrick, J. (2003) [http://www.springerlink.com/content/ju66c6a2734fl20u/ Evaluating Corpora for Named Entity Recognition Using Character-Level Features]. ''Proceeding of the 16th Australian Conference on AI''. Perth, Australia.  


== See also ==
== See also ==

Revision as of 19:35, 31 July 2007

  • Performance measure: F = 2 * Precision * Recall / (Recall + Precision)
  • Precision: percentage of named entities found by the algorithm that are correct
  • Recall: percentage of named entities defined in the corpus that were found by the program
  • Exact calculation of precision and recall is explained in the MUC scoring software
  • Training data: Training section of MUC-7 dataset
  • Testing data: Formal section of MUC-7 dataset


Table of results

System name Short description System type Main publications Software Results (F)
Annotator Human annotator - MUC-7 proceedings - 97.60%
LTG Best MUC-7 participant H Mikheev, Grover and Moens (1998) - 93.39%
Baseline Vocabulary transfer from training to testing S Whitelaw and Patrick (2003) - 58.89%
  • System type: R = hand-crafted rules, S = supervised learning, U = unsupervised learning, H = hybrid


References

Mikheev, A., Grover, C. and Moens, M. (1998). Description of the LTG system used for MUC-7. Proceedings of the Seventh Message Understanding Conference (MUC-7). Fairfax, Virginia.

Whitelaw, C. and Patrick, J. (2003) Evaluating Corpora for Named Entity Recognition Using Character-Level Features. Proceeding of the 16th Australian Conference on AI. Perth, Australia.

See also