Pyspark 拆分列 [英] Pyspark Split Columns
本文介绍了Pyspark 拆分列的处理方法,对大家解决问题具有一定的参考价值,需要的朋友们下面随着小编来一起学习吧!
问题描述
from pyspark.sql import Row, functions as F
row = Row("UK_1","UK_2","Date","Cat",'Combined')
agg = ''
agg = 'Cat'
tdf = (sc.parallelize
([
row(1,1,'12/10/2016',"A",'Water^World'),
row(1,2,None,'A','Sea^Born'),
row(2,1,'14/10/2016','B','Germ^Any'),
row(3,3,'!~2016/2/276','B','Fin^Land'),
row(None,1,'26/09/2016','A','South^Korea'),
row(1,1,'12/10/2016',"A",'North^America'),
row(1,2,None,'A','South^America'),
row(2,1,'14/10/2016','B','New^Zealand'),
row(None,None,'!~2016/2/276','B','South^Africa'),
row(None,1,'26/09/2016','A','Saudi^Arabia')
]).toDF())
cols = F.split(tdf['Combined'], '^')
tdf = tdf.withColumn('column1', cols.getItem(0))
tdf = tdf.withColumn('column2', cols.getItem(1))
tdf.show(truncate = False )
以上是我的示例代码.
出于某种原因,它没有按 ^ 字符拆分列.
For some reason it is not splitting the column by ^ character.
有什么建议吗?
推荐答案
模式是正则表达式,见split;而^
是匹配regex中字符串开头的锚点,要逐字匹配,需要转义:
The pattern is a regular expression, see split; and ^
is an anchor that matches the beginning of string in regex, to match literally, you need to escape it:
cols = F.split(tdf['Combined'], r'\^')
tdf = tdf.withColumn('column1', cols.getItem(0))
tdf = tdf.withColumn('column2', cols.getItem(1))
tdf.show(truncate = False)
+----+----+------------+---+-------------+-------+-------+
|UK_1|UK_2|Date |Cat|Combined |column1|column2|
+----+----+------------+---+-------------+-------+-------+
|1 |1 |12/10/2016 |A |Water^World |Water |World |
|1 |2 |null |A |Sea^Born |Sea |Born |
|2 |1 |14/10/2016 |B |Germ^Any |Germ |Any |
|3 |3 |!~2016/2/276|B |Fin^Land |Fin |Land |
|null|1 |26/09/2016 |A |South^Korea |South |Korea |
|1 |1 |12/10/2016 |A |North^America|North |America|
|1 |2 |null |A |South^America|South |America|
|2 |1 |14/10/2016 |B |New^Zealand |New |Zealand|
|null|null|!~2016/2/276|B |South^Africa |South |Africa |
|null|1 |26/09/2016 |A |Saudi^Arabia |Saudi |Arabia |
+----+----+------------+---+-------------+-------+-------+
这篇关于Pyspark 拆分列的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!
查看全文