hi team! i have packaged my kedro project using th...
# questions
hi team! i have packaged my kedro project using the databricks-cli to be deployed as a databricks job (my first time doing this). after running the task it failed with this error
Copy code
DatasetError: An exception occurred when parsing config for dataset 'my_snowflake_table':
No module named 'snowflake'. Please see the documentation on how to install relevant dependencies for kedro_datasets.snowflake.SnowparkTableDataset:
the pipeline runs fine locally as i have the following installed
Copy code
any insight on this is appreciated!
prior to packaging the kedro project do i need to update the file paths in the catalog to reference where they're stored within dbfs?
It looks like the dependencies needed weren't installed correctly - did you include them in the project's requirements.txt file before packaging?
no i didnt do that. is there a specific command i should use to generate that requirements.txt file? besides pip freeze
@Ahdra Merali that seems to have worked. although the job started it ran for 45+ mins before i canceled it (should not take nearly that long) and the output showed this
Copy code
As an open-source project, we collect usage analytics. 
We cannot see nor store information contained in a Kedro project. 
You can find out more by reading our privacy notice: 
Do you opt into usage analytics?  [y/N]:
i think it got stuck here before actually running the pipeline. how can i avoid this?
looks like someone already asked this in the questions channel, just removed kedro-telemetry from requirements so will see if that works
Hi Nik, you can skip the telemetry prompt by including a
file in your project root and include the following line:
Copy code
consent: <true/false>
👍🏼 1
As for the requirements, it might not be necessary to include everything in you pip freeze but as a rule of thumb, any datasets you use in the catalog should be added to the requirements, this is what it would look like for SnowflakeDataset:
Copy code
got it running - appreciate the help 🙂
np 1
@Ahdra Merali do i need to add anything to the catalog datasets in order for them to be saved to DBFS after each run? i only use 1-2 memory datasets, everything else is a parquet or csv and i thought they'd be saved to DBFS after running but nothing is in the data folders
Could you show me your catalog entries? Is it that the datasets aren't being saved at all, or that they are but in an unexpected location?
This section on uploading data when deploying packaged projects as a Databricks job might also be useful
here is an example of one. and they just arent being saved at all so im sure i just left something out
Copy code
  type: pandas.ParquetDataset
  filepath: data/03_preprocessed/my_df.parquet
      layer: preprocessed
You'll need to set up your catalog to save the data on DBFS - here is an example catalog from the databricks-iris starter project that shows you how to do this Remember to point your conf-source to the right directory as well
got it - so something like this
Copy code
filepath: /dbfs/FileStore/my_product/data/03_preprocessed/my_df.parquet
👍 1