如何从Spark Dataframe中的Row对象获取值? [英] How to get a value from the Row object in Spark Dataframe?
问题描述
对于
averageCount = (wordCountsDF
.groupBy().mean()).head()
我知道
行(平均(count)= 1.6666666666666667)
Row(avg(count)=1.6666666666666667)
但是当我尝试时:
averageCount = (wordCountsDF
.groupBy().mean()).head().getFloat(0)
我收到以下错误:
AttributeError:getFloat -------------------------------------------------- ------------------------- AttributeError Traceback(最近一次调用 最后)在() 1#TODO:替换为适当的代码 ----> 2 averageCount =(wordCountsDF 3 .groupBy().mean()).head().getFloat(0) 4 5次均值计数
AttributeError: getFloat --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) in () 1 # TODO: Replace with appropriate code ----> 2 averageCount = (wordCountsDF 3 .groupBy().mean()).head().getFloat(0) 4 5 print averageCount
/databricks/spark/python/pyspark/sql/types.py(自己,
item)1270引发AttributeError(item)1271
除了ValueError:
-> 1272提高AttributeError(item)1273 1274 def setattr (自身,键,值):
/databricks/spark/python/pyspark/sql/types.py in getattr(self,
item) 1270 raise AttributeError(item) 1271
except ValueError:
-> 1272 raise AttributeError(item) 1273 1274 def setattr(self, key, value):
AttributeError:getFloat
AttributeError: getFloat
我在做什么错了?
推荐答案
我知道了.这将为我返回值:
I figured it out. This will return me the value:
averageCount = (wordCountsDF
.groupBy().mean()).head()[0]
这篇关于如何从Spark Dataframe中的Row对象获取值?的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!