When reloading data from external sources, you may need to refresh existing tables.
Methods:
-
Drop and recreate:
spark.sql("DROP TABLE IF EXISTS example.databricks.my_table") df.write.saveAsTable("example.databricks.my_table") -
Overwrite with schema handling:
df.write.option("overwriteSchema", "true").mode("overwrite").saveAsTable("example.databricks.my_table")
Best Practices:
- Use
DROP TABLEfor inconsistent or manually edited sources (e.g., spreadsheets). - Use
mergeSchemafor additive schema changes. - Always preview with
df.show()before writing.