-
Notifications
You must be signed in to change notification settings - Fork 2.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
when drop a non-Iceberg table , the directory associated with the table was not deleted #11820
Comments
Is this a Bug, or is it by design? |
@lordk911 that is expected. Drop table only remove metadata refs and not the actual data files since 0.14. For actual data file removal, you will need to add You can find the same from the code as well: Entry point for drop table in spark: https://github.com/apache/iceberg/blob/main/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/SparkSessionCatalog.java#L284 iceberg/spark/v3.3/spark/src/main/java/org/apache/iceberg/spark/SparkCatalog.java Line 340 in ab6365d
and the content of
|
Thank you for your reply. I know about this change, but in fact, spark3.3.3 + iceberg 1.3.1, when using SparkSessionCatalog to execute the drop non-Iceberg table, purge keyword is not required. |
I haven't try this on older version of iceberg runtime. The one I currently using is spark 3.5.x + iceberg runtime 1.6.x and drop table without purge won't cleanup datafiles there. I would think that is by designed based on the doc. Wondering if some other bugs is causing data files to get drop in spark3.3.3 + iceberg 1.3.1. |
Apache Iceberg version
1.6.1
Query engine
Spark
Please describe the bug 🐞
spark 3.4.4 with config:
when drop a orc table, the directory associated with the table was not deleted from hdfs, only delete metadata from HMS
Willingness to contribute
The text was updated successfully, but these errors were encountered: