當在簇中處理時,是否可以將驅動程序節點中的spark程序結果的輸出寫入?如何引用執行spark-submit的本地文件系統?
df = sqlContext("hdfs://....")
result = df.groupby('abc','cde').count()
result.write.save("hdfs:...resultfile.parquet", format="parquet") # this works fine
result = result.collect()
with open("<my drivernode local directory>//textfile") as myfile:
myfile.write(result) # I'll convert to python object before writing
有人可以提供一些想法如何引用我給的地方文件系統?