擅长:python、mysql、java
<p>如果行的顺序不重要,这是另一种方法:</p>
<pre><code>from pyspark import SparkContext
sc = SparkContext()
df = sc.parallelize([
{"name":"jack" , "address":"aus", "occupation":"job"},
{"name": "sam", "address":"uk"}
]).toDF()
df = df.na.fill('')
df.show()
+ -+ + +
|address|name|occupation|
+ -+ + +
| aus|jack| job|
| uk| sam| |
+ -+ + +
</code></pre>