Chunking some text with the stanford-nlp

2019-03-25 17:44发布

I'm using the stanford core NLP and I use this line to load some modules to process my text:

props.put("annotators", "tokenize, ssplit, pos, lemma, ner, parse, dcoref");

Is ther a module that i can load to chunks the text?

Or any suggestion with a alterantive way to use the stanford core to chunk some text?

Thank you

2条回答
Root(大扎)
2楼-- · 2019-03-25 18:07

I think the parser output can be used to obtain NP chunks. Take a look at the context-free representation on the Stanford Parser website which provides example output.

查看更多
走好不送
3楼-- · 2019-03-25 18:23

To use chunking with Stanford NLP you can use the following packages:

  • YamCha: SVM-based NP-chunker, also usable for POS tagging, NER, etc. C/C++ open source. Won CoNLL 2000 shared task. (Less automatic than a specialized POS tagger for an end user.)
  • Mark Greenwood's Noun Phrase Chunker: A Java reimplementation of Ramshaw and Marcus (1995).
  • fnTBL: A fast and flexible implementation of Transformation-Based Learning in C++. Includes a POS tagger, but also NP chunking and general chunking models.

Source: http://www-nlp.stanford.edu/links/statnlp.html#NPchunk

查看更多
登录 后发表回答