斯坦福依赖解析器设置和 NLTK [英] Stanford Dependency Parser Setup and NLTK
问题描述
由于danger89 对上一篇文章的回答,我让标准"斯坦福解析器开始工作,Stanford解析器和 NLTK.
So I got the "standard" Stanford Parser to work thanks to danger89's answers to this previous post, Stanford Parser and NLTK.
但是,我现在正在尝试使依赖项解析器正常工作,但上一个链接中突出显示的方法似乎不再有效.这是我的代码:
However, I am now trying to get the dependency parser to work and it seems the method highlighted in the previous link no longer works. Here is my code:
import nltk
import os
java_path = "C:\Program Files\Java\jre1.8.0_51\bin\java.exe"
os.environ['JAVAHOME'] = java_path
from nltk.parse import stanford
os.environ['STANFORD_PARSER'] = 'path/jar'
os.environ['STANFORD_MODELS'] = 'path/jar'
parser = stanford.StanfordDependencyParser(model_path="path/jar/englishPCFG.ser.gz")
sentences = parser.raw_parse_sents(nltk.sent_tokenize("The iPod is expensive but pretty."))
我收到以下错误:'module' 对象没有属性 'StanfordDependencyParser'
I get the following error: 'module' object has no attribute 'StanfordDependencyParser'
我唯一更改的是StanfordParser"中的StanfordDependencyParser".有什么想法可以让它发挥作用吗?
The only thing I changed was "StanfordDependencyParser" from "StanfordParser". Any ideas how I can get this to work?
我还通过导入它来尝试斯坦福神经依赖解析器,如此处的文档所示:http://www.nltk.org/_modules/nltk/parse/stanford.html
I also tried the Stanford Neural Dependency parser by importing it as shown in the documentation here: http://www.nltk.org/_modules/nltk/parse/stanford.html
这个也没用.
对 NLTK 来说很新.在此先感谢您提供任何有用的意见.
Pretty new to NLTK. Thanks in advance for any helpful input.
推荐答案
StanfordDependencyParser
API 是自 NLTK 3.1 版以来创建的新类对象.
The StanfordDependencyParser
API is a new class object created since NLTK version 3.1.
确保您通过 pip 获得最新的 NLTK
Ensure that you have the latest NLTK available either through pip
pip install -U nltk
或通过您的 linux 包管理器,例如:
or through your linux package manager, e.g.:
sudo apt-get python-nltk
或在 Windows 中,下载 https://pypi.python.org/pypi/nltk并安装,它应该会覆盖您以前的 NLTK 版本.
or in windows, download https://pypi.python.org/pypi/nltk and install and it should overwrite your previous NLTK version.
然后您可以使用文档中所示的 API:
Then you can use the API as shown in the documentation:
from nltk.parse.stanford import StanfordDependencyParser
dep_parser=StanfordDependencyParser(model_path="edu/stanford/nlp/models/lexparser/englishPCFG.ser.gz")
print [parse.tree() for parse in dep_parser.raw_parse("The quick brown fox jumps over the lazy dog.")]
[输出]:
[Tree('jumps', [Tree('fox', ['The', 'quick', 'brown']), Tree('dog', ['over', 'the', 'lazy'])])]
(注意:确保你的 jar 和 os.environ
路径正确,在 Windows 中,它是 something\something\some\path
,在unix 它是 something/something/some/path
)
(Note: Make sure you get your path to jar and os.environ
correct, in Windows, it's something\something\some\path
, in unix it's something/something/some/path
)
另见https://github.com/nltk/nltk/wiki/Installing-Third-Party-Software#stanford-tagger-ner-tokenizer-and-parser,当您需要 TL;DR 解决方案时,请参阅
See also https://github.com/nltk/nltk/wiki/Installing-Third-Party-Software#stanford-tagger-ner-tokenizer-and-parser and when you need a TL;DR solution, see https://github.com/alvations/nltk_cli
这篇关于斯坦福依赖解析器设置和 NLTK的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!