对象数据块不是包 com 的成员 [英] object databricks is not a member of package com
问题描述
我正在尝试使用 Zeppelin (HDP 2.6) 在 Spark2 中使用斯坦福 NLP 库.显然,Databricks 为 Spark 的斯坦福 NLP 库构建了包装器.链接:https://github.com/databricks/spark-corenlp
I am trying to use Stanford NLP library in Spark2 using Zeppelin (HDP 2.6). Apparently there is wrapper built by Databricks for the Stanford NLP library for Spark. Link: https://github.com/databricks/spark-corenlp
我已经从 这里下载了上述包装器的 jar 并且还从 这里 下载了斯坦福 NLP jar.然后我在 Zeppelin 的 Spark2 解释器设置中添加了两组 jar 作为依赖项,并重新启动了解释器.
I have downloaded the jar for the above wrapper from here and also downloaded Stanford NLP jars from here. Then I have added both sets of jars as dependencies in Spark2 interpreter settings of Zeppelin and restarted the interpreter.
仍然下面的示例程序给出错误对象数据块不是包com的成员导入 com.databricks.spark.corenlp.functions._"
Still the below sample program gives the error "object databricks is not a member of package com import com.databricks.spark.corenlp.functions._"
import org.apache.spark.sql.functions._
import com.databricks.spark.corenlp.functions._
import sqlContext.implicits._
val input = Seq(
(1, "<xml>Stanford University is located in California. It is a great university.</xml>")
).toDF("id", "text")
val output = input
.select(cleanxml('text).as('doc))
.select(explode(ssplit('doc)).as('sen))
.select('sen, tokenize('sen).as('words), ner('sen).as('nerTags), sentiment('sen).as('sentiment))
output.show(truncate = false)
推荐答案
问题与下载 Databricks corenlp 的 jar 文件有关.我是从这个位置下载的.问题解决了.
The problem was related to downloading the jar file for Databricks corenlp. I downloaded it from this location. Problem solved.
这篇关于对象数据块不是包 com 的成员的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!