How do you handle failures gracefully when writing code in PySpark? Fill in the blanks to complete the statement: ```python _____ # Fill in the blank Spark.read.table("table_name").select("column").write.mode("append").saveAsTable("new_table_name") _____ print(f"query failed") ```