代码之家  ›  专栏  ›  技术社区  ›  Geet

使用R将来自databricks的spark数据帧写入azure数据仓库

  •  1
  • Geet  · 技术社区  · 7 年前

    spark_df.coalesce(1).write.format("com.databricks.spark.csv").option("header", "true").mode("overwrite").save('...path to azure data lake store folder')
    

    1 回复  |  直到 7 年前
        1
  •  1
  •   user10355350    7 年前

    这应该是:

    spark_df %>% 
      coalesce(1L) %>%          # Same as coalesce(1).
      write.df(                 # Generic writer, because there is no csv specific one
        "...path to azure...",  # Path as before 
         source = "csv",        # Since 2.0 you don't need com.databricks 
         mode = "overwrite", 
         header = "true"        # All ... are used as options
      )