什么是此异常火花指 [英] What is means by this exception in spark
问题描述
16/03/30二十三时23分二十秒WARN TaskSetManager:失落任务1.0级0.0(TID 1,10.208.91.144):java.lang.NoClassDefFoundError的:EDU /斯坦福/ NLP /株/ TreebankLanguagePack
在java.lang.Class.getDeclaredFields0(本机方法)
在java.lang.Class.privateGetDeclaredFields(Class.java:2499)
在java.lang.Class.getDeclaredField(Class.java:1951)
在java.io.ObjectStreamClass.getDeclaredSUID(ObjectStreamClass.java:1659)
在java.io.ObjectStreamClass.access $ 700(ObjectStreamClass.java:72)
在java.io.ObjectStreamClass中的$ 2.run(ObjectStreamClass.java:480)
在java.io.ObjectStreamClass中的$ 2.run(ObjectStreamClass.java:468)
在java.security.AccessController.doPrivileged(本机方法)
在java.io.ObjectStreamClass中。(ObjectStreamClass.java:468)
在java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:365)
在java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:602)
在java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622)
在java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
在java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
在java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
在java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
在java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
在java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
在java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
在java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
在java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
在java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
在java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
在java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
在java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
在java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
在java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
在java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
在java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
在java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
在java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
在java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
。在scala.collection.immutable $结肠$ colon.readObject(List.scala:362)
在sun.reflect.NativeMethodAccessorImpl.invoke0(本机方法)
在sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
在sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
在java.lang.reflect.Method.invoke(Method.java:606)
在java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017)
在java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
在java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
在java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
在java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
在java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
在java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
在java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
在java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
在java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
在java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
在java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
在java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
在org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:69)
在org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:95)
在org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:58)
在org.apache.spark.scheduler.Task.run(Task.scala:70)
在org.apache.spark.executor.Executor $ TaskRunner.run(Executor.scala:213)
在java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
在java.util.concurrent.ThreadPoolExecutor中的$ Worker.run(ThreadPoolExecutor.java:615)
在java.lang.Thread.run(Thread.java:745)
抛出java.lang.ClassNotFoundException:引起edu.stanford.nlp.trees.TreebankLanguagePack
在java.net.URLClassLoader的$ 1.run(URLClassLoader.java:366)
在java.net.URLClassLoader的$ 1.run(URLClassLoader.java:355)
在java.security.AccessController.doPrivileged(本机方法)
在java.net.URLClassLoader.findClass(URLClassLoader.java:354)
在java.lang.ClassLoader.loadClass(ClassLoader.java:425)
在java.lang.ClassLoader.loadClass(ClassLoader.java:358)
... 58多个
16/03/30二十三时23分二十秒INFO TaskSetManager:在第一阶段0.0开始任务1.1(TID 2,10.208.91.144,PROCESS_LOCAL,5942字节)
16/03/30二十三点23分20秒INFO TaskSetManager:对执行人10.208.91.144迷失任务0.0舞台0.0(TID 0):java.lang.NoClassDefFoundError的(EDU /斯坦福/ NLP /株/ TreebankLanguagePack)[重复1]
16/03/30二十三点23分20秒INFO TaskSetManager:在第一阶段0.0开始任务0.1(TID 3,10.208.91.144,PROCESS_LOCAL,1435字节)
16/03/30二十三点23分20秒WARN TransportChannelHandler:异常从/10.208.91.144:61788连接
java.io.IOException异常:一个现有的连接被强行关闭远程主机
在sun.nio.ch.SocketDispatcher.read0(本机方法)
在sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:43)
在sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)
在sun.nio.ch.IOUtil.read(IOUtil.java:192)
在sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:380)
在io.netty.buffer.PooledUnsafeDirectByteBuf.setBytes(PooledUnsafeDirectByteBuf.java:311)
在io.netty.buffer.AbstractByteBuf.writeBytes(AbstractByteBuf.java:881)
在io.netty.channel.socket.nio.NioSocketChannel.doReadBytes(NioSocketChannel.java:225)
在io.netty.channel.nio.AbstractNioByteChannel $ NioByteUnsafe.read(AbstractNioByteChannel.java:119)
在io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511)
在io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468)
在io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382)
在io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354)
在io.netty.util.concurrent.SingleThreadEventExecutor $ 2.run(SingleThreadEventExecutor.java:116)
在java.lang.Thread.run(Thread.java:745)
16/03/30二十三点23分20秒错误TaskSchedulerImpl:失落的执行人0 10.208.91.144:远程RPC客户端取消关联
16/03/30二十三点23分20秒INFO TaskSetManager:从使用taskset 0.0 0重新排队任务
16/03/30二十三点23分20秒WARN ReliableDeliverySupervisor:协会远程系统[akka.tcp://sparkExecutor@10.208.91.144:61767]失败,地址选通现为[5000]毫秒。理由是:解除关联。
在第一阶段0.0(TID 2,10.208.91.144)丢失任务1.1:ExecutorLostFailure(0执行人丢失)16/03/30二十三时23分二十秒WARN TaskSetManager
在第一阶段0.0(TID 3,10.208.91.144)丢失任务0.1:ExecutorLostFailure(0执行人丢失)16/03/30二十三时23分二十秒WARN TaskSetManager
16/03/30二十三点23分20秒INFO DAGScheduler:执行人丢失:0(0时代)
16/03/30二十三点23分20秒INFO BlockManagerMasterEndpoint试图从BlockManagerMaster删除执行人0。
16/03/30二十三点23分20秒INFO BlockManagerMasterEndpoint:删除块经理BlockManagerId(0,10.208.91.144,61786)
16/03/30二十三点23分20秒INFO BlockManagerMaster:在removeExecutor成功摘除0
16/03/30二十三点23分20秒INFO $应用程序客户机ClientActor:执行人更新:APP-20160330232314-0002 / 0现在EXITED(命令退出,code 50)
16/03/30二十三点23分20秒INFO SparkDeploySchedulerBackend:执行人APP-20160330232314-0002 /移除0:命令退出,code 50
16/03/30二十三点23分20秒错误SparkDeploySchedulerBackend:要求删除不存在的遗嘱执行人0
16/03/30二十三点23分20秒INFO $应用程序客户机ClientActor:执行人说:APP-20160330232314-0002 / 1工人20160330231130-10.208.91.144-61218(10.208.91.144:61218)4个核心
16/03/30二十三点23分20秒INFO SparkDeploySchedulerBackend:授予执行者ID的应用程序,20160330232314-0002上HOSTPORT 10.208.91.144:61218 / 1 4个核心,512.0 MB RAM
16/03/30二十三点23分20秒INFO $应用程序客户机ClientActor:执行人更新:APP-20160330232314-0002 / 1现在正在运行
16/03/30二十三点23分20秒INFO $应用程序客户机ClientActor:执行人更新:APP-20160330232314-0002 / 1现在是LOADING
16/03/30 23点23分23秒INFO SparkDeploySchedulerBackend:注册执行人:AkkaRpcEndpointRef(演员[akka.tcp://sparkExecutor@10.208.91.144:61815 /用户/执行器#-238863041])ID为1
16/03/30 23点23分23秒INFO TaskSetManager:在第一阶段0.0开始任务0.2(TID 4,10.208.91.144,PROCESS_LOCAL,1435字节)
16/03/30 23点23分23秒INFO TaskSetManager:在第一阶段0.0开始任务1.2(TID 5,10.208.91.144,PROCESS_LOCAL,5942字节)
16/03/30 23点23分24秒INFO BlockManagerMasterEndpoint:注册块管理器10.208.91.144:61834与265.4 MB RAM,BlockManagerId(1 10.208.91.144,61834)
16/03/30 23点23分24秒INFO BlockManagerInfo:新增broadcast_0_piece0在内存上10.208.91.144:61834(尺寸:30.4 KB,自由:265.4 MB)
16/03/30 23点23分24秒INFO TaskSetManager:对执行人10.208.91.144迷失任务1.2级0.0(TID 5):java.lang.NoClassDefFoundError的(EDU /斯坦福/ NLP /株/ TreebankLanguagePack)[重复2]
16/03/30 23点23分24秒INFO TaskSetManager:在第一阶段0.0开始任务1.3(TID 6,10.208.91.144,PROCESS_LOCAL,5942字节)
16/03/30 23点23分24秒INFO TaskSetManager:对执行人10.208.91.144迷失任务0.2舞台0.0(TID 4):java.lang.NoClassDefFoundError的(EDU /斯坦福/ NLP /株/ TreebankLanguagePack)[重复3]
16/03/30 23点23分24秒INFO TaskSetManager:在第一阶段0.0开始任务0.3(TID 7,10.208.91.144,PROCESS_LOCAL,1435字节)
16/03/30二十三点23分25秒WARN TransportChannelHandler:异常从/10.208.91.144:61835连接
java.io.IOException异常:一个现有的连接被强行关闭远程主机
在sun.nio.ch.SocketDispatcher.read0(本机方法)
在sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:43)
在sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)
在sun.nio.ch.IOUtil.read(IOUtil.java:192)
在sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:380)
在io.netty.buffer.PooledUnsafeDirectByteBuf.setBytes(PooledUnsafeDirectByteBuf.java:311)
在io.netty.buffer.AbstractByteBuf.writeBytes(AbstractByteBuf.java:881)
在io.netty.channel.socket.nio.NioSocketChannel.doReadBytes(NioSocketChannel.java:225)
在io.netty.channel.nio.AbstractNioByteChannel $ NioByteUnsafe.read(AbstractNioByteChannel.java:119)
在io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511)
在io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468)
在io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382)
在io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354)
在io.netty.util.concurrent.SingleThreadEventExecutor $ 2.run(SingleThreadEventExecutor.java:116)
在java.lang.Thread.run(Thread.java:745)
16/03/30二十三点23分25秒错误TaskSchedulerImpl:失落的遗嘱执行人1上10.208.91.144:远程RPC客户端取消关联
16/03/30二十三点23分25秒INFO TaskSetManager:从使用taskset 0.0 1重新排队任务
在第一阶段0.0(TID 7,10.208.91.144)丢失任务0.3:ExecutorLostFailure(遗嘱执行人1赔)16/03/30 23点23分25秒WARN TaskSetManager
16/03/30二十三点23分25秒WARN ReliableDeliverySupervisor:协会远程系统[akka.tcp://sparkExecutor@10.208.91.144:61815]失败,地址选通现为[5000]毫秒。理由是:解除关联。
16/03/30二十三点23分25秒错误TaskSetManager:任务0级0.0失败了4倍;中止工作
在第一阶段0.0(TID 6,10.208.91.144)丢失任务1.3:ExecutorLostFailure(遗嘱执行人1赔)16/03/30 23点23分25秒WARN TaskSetManager
16/03/30二十三点23分25秒INFO TaskSchedulerImpl:删除taskset的0.0,其任务已全部建成后,从池
16/03/30二十三点23分25秒INFO TaskSchedulerImpl:取消0期
16/03/30二十三点23分25秒INFO DAGScheduler:ResultStage 0(saveAsTextFile在Polarity.java:62)未能8.085小号
16/03/30二十三点23分25秒INFO DAGScheduler:作业0失败:saveAsTextFile在Polarity.java:62,拿了8.447334小号
16/03/30二十三点23分25秒INFO DAGScheduler:执行人丢失:1(1时代)
org.apache.spark.SparkException:作业已中止由于舞台失败:任务0级0.0失败了4次,最近一次故障:在第一阶段0.0迷失任务0.3(TID 7,10.208.91.144):ExecutorLostFailure(遗嘱执行人1赔)
驱动程序堆栈跟踪:
在org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1266)
在org.apache.spark.scheduler.DAGScheduler $$ anonfun $ abortStage $ 1.适用(DAGScheduler.scala:1257)
在org.apache.spark.scheduler.DAGScheduler $$ anonfun $ abortStage $ 1.适用(DAGScheduler.scala:1256)
在scala.collection.mutable.ResizableArray $ class.foreach(ResizableArray.scala:59)
在scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
在org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1256)
在org.apache.spark.scheduler.DAGScheduler $$ anonfun $ handleTaskSetFailed $ 1.适用(DAGScheduler.scala:730)
在org.apache.spark.scheduler.DAGScheduler $$ anonfun $ handleTaskSetFailed $ 1.适用(DAGScheduler.scala:730)
在scala.Option.foreach(Option.scala:236)
在org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:730)
在org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1450)
在org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1411)
在org.apache.spark.util.EventLoop $$不久$ 1.run(EventLoop.scala:48)
异常线程mainjava.io.FileNotFoundException:G:\\工作区\\ DSS \\文本文件\\兼职00000(该系统找不到指定的文件)
在java.io.FileInputStream.open0(本机方法)
在java.io.FileInputStream.open(FileInputStream.java:195)
在java.io.FileInputStream中。(FileInputStream.java:138)
在java.io.FileInputStream中。(FileInputStream.java:93)
在java.io.FileReader(FileReader.java:58)
在com.Polarity.read(Polarity.java:94)
在com.Polarity.main(Polarity.java:288)
16/03/30二十三点23分25秒INFO BlockManagerMasterEndpoint试图从BlockManagerMaster删除执行人1。
16/03/30二十三点23分25秒INFO BlockManagerMasterEndpoint:删除块经理BlockManagerId(1 10.208.91.144,61834)
16/03/30二十三点23分25秒INFO BlockManagerMaster:在removeExecutor成功取出1
16/03/30二十三点23分25秒INFO $应用程序客户机ClientActor:执行人更新:APP-20160330232314-0002 / 1现在EXITED(命令退出,code 50)
16/03/30二十三点23分25秒INFO SparkDeploySchedulerBackend:执行人APP-20160330232314-0002 / 1中删除:命令退出,code 50
16/03/30二十三点23分25秒错误SparkDeploySchedulerBackend:要求删除不存在的遗嘱执行人1
16/03/30二十三点23分25秒INFO $应用程序客户机ClientActor:执行人说:APP-20160330232314-0002 / 2工人20160330231130-10.208.91.144-61218(10.208.91.144:61218)4个核心
16/03/30二十三点23分25秒INFO SparkDeploySchedulerBackend:授予执行者ID的应用程序,20160330232314-0002上HOSTPORT 10.208.91.144:61218 / 2 4个核心,512.0 MB RAM
16/03/30二十三点23分25秒INFO SparkContext:从关闭挂钩调用的stop()
16/03/30二十三点23分25秒INFO $应用程序客户机ClientActor:执行人更新:APP-20160330232314-0002 / 2现在正在运行
16/03/30二十三点23分25秒INFO $应用程序客户机ClientActor:执行人更新:APP-20160330232314-0002 / 2现在LOADING
16/03/30二十三点23分25秒INFO SparkUI:在停止星火web用户界面 http://10.208.91.144:4040一>
16/03/30二十三点23分25秒INFO DAGScheduler:停止DAGScheduler
16/03/30二十三点23分25秒INFO SparkDeploySchedulerBackend:关闭所有执行人
16/03/30二十三点23分25秒INFO SparkDeploySchedulerBackend:要求每个执行人关闭
16/03/30二十三点23分25秒INFO MapOutputTrackerMasterEndpoint:MapOutputTrackerMasterEndpoint停了!
16/03/30二十三点23分25秒INFO utils的:PATH = C:\\Users\\Developer\\AppData\\Local\\Temp\\spark-618a9039-a9d4-4fb2-bdc8-408d1e7f3c0e\\blockmgr-4436a6d6-fca4-4190-ac2a-48c8ebd4e7db,已经present作为根删除。
16/03/30二十三点23分25秒INFO MemoryStore的:MemoryStore的清除
16/03/30二十三点23分25秒INFO BlockManager:BlockManager停止
16/03/30二十三点23分25秒INFO BlockManagerMaster:BlockManagerMaster停止
16/03/30二十三点23分25秒INFO OutputCommitCoordinator $ OutputCommitCoordinatorEndpoint:OutputCommitCoordinator停了!
16/03/30二十三点23分25秒INFO RemoteActorRefProvider $ RemotingTerminator:关闭远程后台程序。
16/03/30二十三点23分25秒INFO SparkContext:已成功停止SparkContext
16/03/30二十三点23分25秒INFO RemoteActorRefProvider $ RemotingTerminator:远程后台程序关闭;有冲洗远程传输进行。
16/03/30二十三点23分25秒INFO utils的:所谓的关闭挂钩
16/03/30二十三点23分25秒INFO utils的:删除目录C:\\用户\\开发\\应用程序数据\\本地的\\ Temp \\火花618a9039-a9d4-4fb2-bdc8-408d1e7f3c0e
这是我的pom.xml
<项目的xmlns =http://maven.apache.org/POM/4.0.0的xmlns:XSI =http://www.w3.org/2001 / XML模式实例XSI:的schemaLocation =http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd\">
< modelVersion> 4.0.0< / modelVersion>
<&的groupId GT;五规划< /的groupId>
<&的artifactId GT; DSS< / artifactId的>
<&版GT; 0.0.1-SNAPSHOT< /版本>
<建立>
<&插件GT;
<&插件GT;
<&的groupId GT; org.apache.maven.plugins< /的groupId>
<&的artifactId GT; Maven的编译器插件< / artifactId的>
<&版GT; 2.3.2< /版本>
<结构>
<信源> 1.8 LT; /源>
<目标> 1.8 LT; /目标与GT;
< /结构>
< /插件>
<&插件GT;
<&的groupId GT; org.apache.maven.plugins< /的groupId>
<&的artifactId GT;的maven-JAR-插件< / artifactId的>
<结构>
<归档和GT;
<&舱单GT;
< addClasspath>真< / addClasspath>
<类路径preFIX>的lib /< /类路径preFIX>
< mainClass> com.Polarity< / mainClass>
< /清单>
< /存档>
< /结构>
< /插件>
<&插件GT;
<&的groupId GT; org.apache.maven.plugins< /的groupId>
<&的artifactId GT; Maven的编译器插件< / artifactId的>
<&执行GT;
<执行与GT;
<阶段>&编译LT; /阶段>
<目标>
<&目标GT;&编译LT; /目标>
< /目标>
< /执行>
< /处决>
< /插件>
< /插件>
< /构建> <依赖和GT; <! - 进口星火 - >
<&依赖性GT;
<&的groupId GT; org.apache.spark< /的groupId>
<&的artifactId GT;火花core_2.10< / artifactId的>
<&版GT; 1.4.0< /版本>
< /依赖性>
<&依赖性GT;
<&的groupId GT; org.apache.spark< /的groupId>
<&的artifactId GT;火花sql_2.10< / artifactId的>
<&版GT; 1.4.0< /版本>
< /依赖性>
<&依赖性GT;
<&的groupId GT; org.springframework< /的groupId>
<&的artifactId GT;弹簧芯LT; / artifactId的>
<&版GT; 2.5 LT; /版本>
< /依赖性>
<&依赖性GT;
<&的groupId GT; org.apache.spark< /的groupId>
<&的artifactId GT;火花sql_2.10< / artifactId的>
<&版GT; 1.3.1< /版本>
< /依赖性>
<&依赖性GT;
<&的groupId GT;夹爪LT; /的groupId>
<&的artifactId GT;夹爪LT; / artifactId的>
<&版GT; 1.2 LT; /版本>
<&型GT;&罐子LT; /类型>
<&范围GT;系统和LT; /&范围GT;
&LT; Systempath下&GT; G:/Workspace/DSS/lib/jaws-bin.jar< / Systempath下&GT;
&LT; /依赖性&GT;
&LT;&依赖性GT;
&LT;&的groupId GT;的commons-logging&LT; /的groupId&GT;
&LT;&的artifactId GT;的commons-logging&LT; / artifactId的&GT;
&LT;&版GT; 1.1.3&LT; /版本&GT;
&LT; /依赖性&GT;
&LT;&依赖性GT;
&LT;&的groupId GT; edu.stanford.nlp&LT; /的groupId&GT;
&LT;&的artifactId GT;斯坦福大学corenlp&LT; / artifactId的&GT;
&LT;&版GT; 3.5.2&LT; /版本&GT;
&LT; /依赖性&GT;
&LT;&依赖性GT;
&LT;&的groupId GT; edu.stanford.nlp&LT; /的groupId&GT;
&LT;&的artifactId GT;斯坦福大学的解析器&LT; / artifactId的&GT;
&LT;&版GT; 2.0.2&LT; /版本&GT;
&LT; /依赖性&GT;
&LT;&依赖性GT;
&LT;&的groupId GT; edu.stanford.nlp&LT; /的groupId&GT;
&LT;&的artifactId GT;斯坦福大学corenlp&LT; / artifactId的&GT;
&LT;&版GT; 3.5.0&LT; /版本&GT;
&LT;分类器GT;型号和LT; /分类&GT;
&LT; /依赖性&GT;
&LT;&依赖性GT;
&LT;&的groupId GT; com.google code.json简单的&LT; /的groupId&GT;
&LT;&的artifactId GT; JSON-简单&LT; / artifactId的&GT;
&LT;&版GT; 1.1&LT; /版本&GT;
&LT; /依赖性&GT;
&LT;&依赖性GT;
&LT;&的groupId GT;的JUnit&LT; /的groupId&GT;
&LT;&的artifactId GT;的JUnit&LT; / artifactId的&GT;
&LT;&版GT;&4.11 LT; /版本&GT;
&LT;&范围GT;试验&LT; /&范围GT;
&LT; /依赖性&GT;
&LT; /依赖和GT;
&LT;性状&gt;
&LT; project.build.sourceEncoding&GT; UTF-8&LT; /project.build.sourceEncoding>
&LT; /性状&gt;&LT; /项目&GT;
堆栈跟踪的第一行告诉你问题是什么。
java.lang.NoClassDefFoundError的:EDU /斯坦福/ NLP /株/ TreebankLanguagePack在
如果您搜索谷歌或为计算器的NoClassDefFoundError
,你会发现描述如何解决这个问题的文章。这不是具体到Apache的火花。这是一个普通的Java问题。在Java类路径中被错误地设置了所有的可能性---你错过classpath中的罐子的文件,有一个罐子的版本错误,或者没有包含在类路径中的所有类文件。
您可能需要找出其中罐子包含类 edu.stanford.nlp.trees.TreebankLanguagePack
。谷歌搜索该类名称,表明它是斯坦福parser.jar
您缺少。您可能会丢失其他罐子了。
更新:现在你已经发布你的Maven的配置。我想你已经指定了一个旧版本斯坦福parser.jar中不包含 TreebankLanguagePack
类。试试这个:
&LT;&依赖性GT;
&LT;&的groupId GT; edu.stanford.nlp&LT; /的groupId&GT;
&LT;&的artifactId GT;斯坦福大学的解析器&LT; / artifactId的&GT;
&LT;&版GT; 3.6.0&LT; /版本&GT;
&LT; /依赖性&GT;
16/03/30 23:23:20 WARN TaskSetManager: Lost task 1.0 in stage 0.0 (TID 1, 10.208.91.144): java.lang.NoClassDefFoundError: edu/stanford/nlp/trees/TreebankLanguagePack at java.lang.Class.getDeclaredFields0(Native Method) at java.lang.Class.privateGetDeclaredFields(Class.java:2499) at java.lang.Class.getDeclaredField(Class.java:1951) at java.io.ObjectStreamClass.getDeclaredSUID(ObjectStreamClass.java:1659) at java.io.ObjectStreamClass.access$700(ObjectStreamClass.java:72) at java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:480) at java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:468) at java.security.AccessController.doPrivileged(Native Method) at java.io.ObjectStreamClass.(ObjectStreamClass.java:468) at java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:365) at java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:602) at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622) at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) at scala.collection.immutable.$colon$colon.readObject(List.scala:362) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) at org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:69) at org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:95) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:58) at org.apache.spark.scheduler.Task.run(Task.scala:70) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) Caused by: java.lang.ClassNotFoundException: edu.stanford.nlp.trees.TreebankLanguagePack at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) ... 58 more
16/03/30 23:23:20 INFO TaskSetManager: Starting task 1.1 in stage 0.0 (TID 2, 10.208.91.144, PROCESS_LOCAL, 5942 bytes) 16/03/30 23:23:20 INFO TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0) on executor 10.208.91.144: java.lang.NoClassDefFoundError (edu/stanford/nlp/trees/TreebankLanguagePack) [duplicate 1] 16/03/30 23:23:20 INFO TaskSetManager: Starting task 0.1 in stage 0.0 (TID 3, 10.208.91.144, PROCESS_LOCAL, 1435 bytes) 16/03/30 23:23:20 WARN TransportChannelHandler: Exception in connection from /10.208.91.144:61788 java.io.IOException: An existing connection was forcibly closed by the remote host at sun.nio.ch.SocketDispatcher.read0(Native Method) at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:43) at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223) at sun.nio.ch.IOUtil.read(IOUtil.java:192) at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:380) at io.netty.buffer.PooledUnsafeDirectByteBuf.setBytes(PooledUnsafeDirectByteBuf.java:311) at io.netty.buffer.AbstractByteBuf.writeBytes(AbstractByteBuf.java:881) at io.netty.channel.socket.nio.NioSocketChannel.doReadBytes(NioSocketChannel.java:225) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:119) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:116) at java.lang.Thread.run(Thread.java:745) 16/03/30 23:23:20 ERROR TaskSchedulerImpl: Lost executor 0 on 10.208.91.144: remote Rpc client disassociated 16/03/30 23:23:20 INFO TaskSetManager: Re-queueing tasks for 0 from TaskSet 0.0 16/03/30 23:23:20 WARN ReliableDeliverySupervisor: Association with remote system [akka.tcp://sparkExecutor@10.208.91.144:61767] has failed, address is now gated for [5000] ms. Reason is: [Disassociated]. 16/03/30 23:23:20 WARN TaskSetManager: Lost task 1.1 in stage 0.0 (TID 2, 10.208.91.144): ExecutorLostFailure (executor 0 lost) 16/03/30 23:23:20 WARN TaskSetManager: Lost task 0.1 in stage 0.0 (TID 3, 10.208.91.144): ExecutorLostFailure (executor 0 lost) 16/03/30 23:23:20 INFO DAGScheduler: Executor lost: 0 (epoch 0) 16/03/30 23:23:20 INFO BlockManagerMasterEndpoint: Trying to remove executor 0 from BlockManagerMaster. 16/03/30 23:23:20 INFO BlockManagerMasterEndpoint: Removing block manager BlockManagerId(0, 10.208.91.144, 61786) 16/03/30 23:23:20 INFO BlockManagerMaster: Removed 0 successfully in removeExecutor 16/03/30 23:23:20 INFO AppClient$ClientActor: Executor updated: app-20160330232314-0002/0 is now EXITED (Command exited with code 50) 16/03/30 23:23:20 INFO SparkDeploySchedulerBackend: Executor app-20160330232314-0002/0 removed: Command exited with code 50 16/03/30 23:23:20 ERROR SparkDeploySchedulerBackend: Asked to remove non-existent executor 0 16/03/30 23:23:20 INFO AppClient$ClientActor: Executor added: app-20160330232314-0002/1 on worker-20160330231130-10.208.91.144-61218 (10.208.91.144:61218) with 4 cores 16/03/30 23:23:20 INFO SparkDeploySchedulerBackend: Granted executor ID app-20160330232314-0002/1 on hostPort 10.208.91.144:61218 with 4 cores, 512.0 MB RAM 16/03/30 23:23:20 INFO AppClient$ClientActor: Executor updated: app-20160330232314-0002/1 is now RUNNING 16/03/30 23:23:20 INFO AppClient$ClientActor: Executor updated: app-20160330232314-0002/1 is now LOADING 16/03/30 23:23:23 INFO SparkDeploySchedulerBackend: Registered executor: AkkaRpcEndpointRef(Actor[akka.tcp://sparkExecutor@10.208.91.144:61815/user/Executor#-238863041]) with ID 1 16/03/30 23:23:23 INFO TaskSetManager: Starting task 0.2 in stage 0.0 (TID 4, 10.208.91.144, PROCESS_LOCAL, 1435 bytes) 16/03/30 23:23:23 INFO TaskSetManager: Starting task 1.2 in stage 0.0 (TID 5, 10.208.91.144, PROCESS_LOCAL, 5942 bytes) 16/03/30 23:23:24 INFO BlockManagerMasterEndpoint: Registering block manager 10.208.91.144:61834 with 265.4 MB RAM, BlockManagerId(1, 10.208.91.144, 61834) 16/03/30 23:23:24 INFO BlockManagerInfo: Added broadcast_0_piece0 in memory on 10.208.91.144:61834 (size: 30.4 KB, free: 265.4 MB) 16/03/30 23:23:24 INFO TaskSetManager: Lost task 1.2 in stage 0.0 (TID 5) on executor 10.208.91.144: java.lang.NoClassDefFoundError (edu/stanford/nlp/trees/TreebankLanguagePack) [duplicate 2] 16/03/30 23:23:24 INFO TaskSetManager: Starting task 1.3 in stage 0.0 (TID 6, 10.208.91.144, PROCESS_LOCAL, 5942 bytes) 16/03/30 23:23:24 INFO TaskSetManager: Lost task 0.2 in stage 0.0 (TID 4) on executor 10.208.91.144: java.lang.NoClassDefFoundError (edu/stanford/nlp/trees/TreebankLanguagePack) [duplicate 3] 16/03/30 23:23:24 INFO TaskSetManager: Starting task 0.3 in stage 0.0 (TID 7, 10.208.91.144, PROCESS_LOCAL, 1435 bytes) 16/03/30 23:23:25 WARN TransportChannelHandler: Exception in connection from /10.208.91.144:61835 java.io.IOException: An existing connection was forcibly closed by the remote host at sun.nio.ch.SocketDispatcher.read0(Native Method) at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:43) at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223) at sun.nio.ch.IOUtil.read(IOUtil.java:192) at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:380) at io.netty.buffer.PooledUnsafeDirectByteBuf.setBytes(PooledUnsafeDirectByteBuf.java:311) at io.netty.buffer.AbstractByteBuf.writeBytes(AbstractByteBuf.java:881) at io.netty.channel.socket.nio.NioSocketChannel.doReadBytes(NioSocketChannel.java:225) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:119) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354) at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:116) at java.lang.Thread.run(Thread.java:745) 16/03/30 23:23:25 ERROR TaskSchedulerImpl: Lost executor 1 on 10.208.91.144: remote Rpc client disassociated 16/03/30 23:23:25 INFO TaskSetManager: Re-queueing tasks for 1 from TaskSet 0.0 16/03/30 23:23:25 WARN TaskSetManager: Lost task 0.3 in stage 0.0 (TID 7, 10.208.91.144): ExecutorLostFailure (executor 1 lost) 16/03/30 23:23:25 WARN ReliableDeliverySupervisor: Association with remote system [akka.tcp://sparkExecutor@10.208.91.144:61815] has failed, address is now gated for [5000] ms. Reason is: [Disassociated]. 16/03/30 23:23:25 ERROR TaskSetManager: Task 0 in stage 0.0 failed 4 times; aborting job 16/03/30 23:23:25 WARN TaskSetManager: Lost task 1.3 in stage 0.0 (TID 6, 10.208.91.144): ExecutorLostFailure (executor 1 lost) 16/03/30 23:23:25 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose tasks have all completed, from pool 16/03/30 23:23:25 INFO TaskSchedulerImpl: Cancelling stage 0 16/03/30 23:23:25 INFO DAGScheduler: ResultStage 0 (saveAsTextFile at Polarity.java:62) failed in 8.085 s 16/03/30 23:23:25 INFO DAGScheduler: Job 0 failed: saveAsTextFile at Polarity.java:62, took 8.447334 s 16/03/30 23:23:25 INFO DAGScheduler: Executor lost: 1 (epoch 1) org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 7, 10.208.91.144): ExecutorLostFailure (executor 1 lost) Driver stacktrace: at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1266) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1257) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1256) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1256) at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:730) at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:730) at scala.Option.foreach(Option.scala:236) at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:730) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1450) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1411) at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) Exception in thread "main" java.io.FileNotFoundException: G:\Workspace\DSS\TextFile\part-00000 (The system cannot find the file specified) at java.io.FileInputStream.open0(Native Method) at java.io.FileInputStream.open(FileInputStream.java:195) at java.io.FileInputStream.(FileInputStream.java:138) at java.io.FileInputStream.(FileInputStream.java:93) at java.io.FileReader.(FileReader.java:58) at com.Polarity.read(Polarity.java:94) at com.Polarity.main(Polarity.java:288) 16/03/30 23:23:25 INFO BlockManagerMasterEndpoint: Trying to remove executor 1 from BlockManagerMaster. 16/03/30 23:23:25 INFO BlockManagerMasterEndpoint: Removing block manager BlockManagerId(1, 10.208.91.144, 61834) 16/03/30 23:23:25 INFO BlockManagerMaster: Removed 1 successfully in removeExecutor 16/03/30 23:23:25 INFO AppClient$ClientActor: Executor updated: app-20160330232314-0002/1 is now EXITED (Command exited with code 50) 16/03/30 23:23:25 INFO SparkDeploySchedulerBackend: Executor app-20160330232314-0002/1 removed: Command exited with code 50 16/03/30 23:23:25 ERROR SparkDeploySchedulerBackend: Asked to remove non-existent executor 1 16/03/30 23:23:25 INFO AppClient$ClientActor: Executor added: app-20160330232314-0002/2 on worker-20160330231130-10.208.91.144-61218 (10.208.91.144:61218) with 4 cores 16/03/30 23:23:25 INFO SparkDeploySchedulerBackend: Granted executor ID app-20160330232314-0002/2 on hostPort 10.208.91.144:61218 with 4 cores, 512.0 MB RAM 16/03/30 23:23:25 INFO SparkContext: Invoking stop() from shutdown hook 16/03/30 23:23:25 INFO AppClient$ClientActor: Executor updated: app-20160330232314-0002/2 is now RUNNING 16/03/30 23:23:25 INFO AppClient$ClientActor: Executor updated: app-20160330232314-0002/2 is now LOADING 16/03/30 23:23:25 INFO SparkUI: Stopped Spark web UI at http://10.208.91.144:4040 16/03/30 23:23:25 INFO DAGScheduler: Stopping DAGScheduler 16/03/30 23:23:25 INFO SparkDeploySchedulerBackend: Shutting down all executors 16/03/30 23:23:25 INFO SparkDeploySchedulerBackend: Asking each executor to shut down 16/03/30 23:23:25 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped! 16/03/30 23:23:25 INFO Utils: path = C:\Users\Developer\AppData\Local\Temp\spark-618a9039-a9d4-4fb2-bdc8-408d1e7f3c0e\blockmgr-4436a6d6-fca4-4190-ac2a-48c8ebd4e7db, already present as root for deletion. 16/03/30 23:23:25 INFO MemoryStore: MemoryStore cleared 16/03/30 23:23:25 INFO BlockManager: BlockManager stopped 16/03/30 23:23:25 INFO BlockManagerMaster: BlockManagerMaster stopped 16/03/30 23:23:25 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped! 16/03/30 23:23:25 INFO RemoteActorRefProvider$RemotingTerminator: Shutting down remote daemon. 16/03/30 23:23:25 INFO SparkContext: Successfully stopped SparkContext 16/03/30 23:23:25 INFO RemoteActorRefProvider$RemotingTerminator: Remote daemon shut down; proceeding with flushing remote transports. 16/03/30 23:23:25 INFO Utils: Shutdown hook called 16/03/30 23:23:25 INFO Utils: Deleting directory C:\Users\Developer\AppData\Local\Temp\spark-618a9039-a9d4-4fb2-bdc8-408d1e7f3c0e this is my pom.xml
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
<modelVersion>4.0.0</modelVersion>
<groupId>Fyp</groupId>
<artifactId>DSS</artifactId>
<version>0.0.1-SNAPSHOT</version>
<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<version>2.3.2</version>
<configuration>
<source>1.8</source>
<target>1.8</target>
</configuration>
</plugin>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-jar-plugin</artifactId>
<configuration>
<archive>
<manifest>
<addClasspath>true</addClasspath>
<classpathPrefix>lib/</classpathPrefix>
<mainClass>com.Polarity</mainClass>
</manifest>
</archive>
</configuration>
</plugin>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<executions>
<execution>
<phase>compile</phase>
<goals>
<goal>compile</goal>
</goals>
</execution>
</executions>
</plugin>
</plugins>
</build>
<dependencies>
<!-- Import Spark -->
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-core_2.10</artifactId>
<version>1.4.0</version>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_2.10</artifactId>
<version>1.4.0</version>
</dependency>
<dependency>
<groupId>org.springframework</groupId>
<artifactId>spring-core</artifactId>
<version>2.5</version>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-sql_2.10</artifactId>
<version>1.3.1</version>
</dependency>
<dependency>
<groupId>jaws</groupId>
<artifactId>jaws</artifactId>
<version>1.2</version>
<type>jar</type>
<scope>system</scope>
<systemPath>G:/Workspace/DSS/lib/jaws-bin.jar</systemPath>
</dependency>
<dependency>
<groupId>commons-logging</groupId>
<artifactId>commons-logging</artifactId>
<version>1.1.3</version>
</dependency>
<dependency>
<groupId>edu.stanford.nlp</groupId>
<artifactId>stanford-corenlp</artifactId>
<version>3.5.2</version>
</dependency>
<dependency>
<groupId>edu.stanford.nlp</groupId>
<artifactId>stanford-parser</artifactId>
<version>2.0.2</version>
</dependency>
<dependency>
<groupId>edu.stanford.nlp</groupId>
<artifactId>stanford-corenlp</artifactId>
<version>3.5.0</version>
<classifier>models</classifier>
</dependency>
<dependency>
<groupId>com.googlecode.json-simple</groupId>
<artifactId>json-simple</artifactId>
<version>1.1</version>
</dependency>
<dependency>
<groupId>junit</groupId>
<artifactId>junit</artifactId>
<version>4.11</version>
<scope>test</scope>
</dependency>
</dependencies>
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
</properties>
</project>
The first line of the stack trace tells you what the problem is.
java.lang.NoClassDefFoundError: edu/stanford/nlp/trees/TreebankLanguagePack at
If you search Google or StackOverflow for NoClassDefFoundError
, you'll find articles describing how to fix this problem. It's not specific to Apache Spark. It's a general Java problem. In all likelihood your Java classpath is incorrectly set up---you're missing a "jar" file on the classpath, have the wrong version of a jar or have not included all the class files on the classpath.
You probably need to work out which "jar" contains the class edu.stanford.nlp.trees.TreebankLanguagePack
. Searching Google for that class name, suggests it's the stanford-parser.jar
that you are missing. You may be missing other jars too.
UPDATE: Now you've posted your maven config. I think you have specified an old version of stanford-parser.jar that does not contain the TreebankLanguagePack
class. Try this instead:
<dependency>
<groupId>edu.stanford.nlp</groupId>
<artifactId>stanford-parser</artifactId>
<version>3.6.0</version>
</dependency>
这篇关于什么是此异常火花指的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!