Júlio Resende
09/15/2025, 5:00 PMElena Khaustova
09/16/2025, 9:40 AMabfs:// or abfss://, Spark defers to the Azure Hadoop connector to handle authentication.
• That connector ignores per-writer .options(...) for authentication and only checks Hadoop
• So unless you’ve already set these globally on the Spark session, .save() will fail to authenticate, even if you passed credentials into the Kedro dataset
• For some formats (e.g., JDBC, S3 connectors), Spark allows passing authentication tokens directly as reader options. That’s why Kedro’s SparkDataSet supports merging credentials into .load().Elena Khaustova
09/16/2025, 9:42 AMDataFrameReader.
Writing: those same credentials don’t get passed to DataFrameWriter. Spark tries to resolve the ABFS path and falls back to Hadoop configs.Nok Lam Chan
09/16/2025, 1:18 PMfsspec based).
You can still have multiple spark.yml configuration to keep different set of Spark credentials - though it's not as granular as a dataset level credentials.Júlio Resende
09/16/2025, 1:42 PM.options(...) method