Tooba Mukhtar
12/12/2022, 3:30 PMYetunde
12/12/2022, 3:51 PMYetunde
12/12/2022, 3:53 PMAnirudh Dahiya
12/13/2022, 11:40 AMOlivia Lihn
12/13/2022, 9:30 PMspark.conf.set("fs.azure.account.auth.type.<storage-account>.<http://dfs.core.windows.net|dfs.core.windows.net>", "SAS")
spark.conf.set("fs.azure.sas.token.provider.type.<storage-account>.<http://dfs.core.windows.net|dfs.core.windows.net>", "org.apache.hadoop.fs.azurebfs.sas.FixedSASTokenProvider")
spark.conf.set("fs.azure.sas.fixed.token.<storage-account>.<http://dfs.core.windows.net|dfs.core.windows.net>", "<token>")
how should i set this credentials in the catalog? Do i need to create a custom dataset?Thaiza
12/13/2022, 11:23 PMRickard Ström
12/14/2022, 6:06 PMJordan
12/15/2022, 10:55 AMREADME.md
files of my project. How should I document the type
of a partitioned dataset? Those function outputs need to be of the form dict[str, <type>]
, but when the dataset is loaded back it’s going to be dict[str, Callable[[], <type>]]
Simon Myway
12/15/2022, 11:42 AMBalazs Konig
12/15/2022, 1:27 PMdata_type_a_1.csv
, d`ata_type_a_2.csv`… and I want to read those in together - tried to simply put *
in the filename_suffix, eg. "data_type_a_*.csv"
, but that’s not working, so I’m definitely missing something simple here 😅Anastasiia
12/15/2022, 2:01 PMhttps://kedro.readthedocs.io/en/0.17.7/_images/experiment-tracking_demo_small.gif▾
Jaakko
12/15/2022, 6:59 PMMaurits
12/15/2022, 8:06 PMSlackbot
12/19/2022, 9:00 AMSzymon Czop
12/19/2022, 11:22 AMMaurits
12/19/2022, 11:24 AMkedro-great
seems outdated and generates an error for me while running kedro great init
ImportError: cannot import name 'BatchMarkers' from 'great_expectations.datasource.types'
Luiz Henrique Aguiar
12/19/2022, 1:15 PMsc = SparkContext(conf=spark_conf, appName="Kedro")
_spark_session = (
SparkSession.builder
.appName(context._package_name)
.enableHiveSupport()
.master("local[*,4]")
.getOrCreate()
)
_spark_session.sparkContext.setLogLevel("WARN")
Error:
Error: py4j.protocol.Py4JJavaError: An error occurred while calling None.org.apache.spark.api.java.JavaSparkContext. : org.apache.spark.SparkException: In Databricks, developers should utilize the shared SparkContext instead of creating one using the constructor. In Scala and Python notebooks, the shared context can be accessed as sc. When running a job, you can access the shared context by calling SparkContext.getOrCreate(). The other SparkContext was created at: CallSite(SparkContext at DatabricksILoop.scala:353,org.apache.spark.SparkContext.(SparkContext.scala:114)
I've tried to delete the Hook and make the Spark settings directly in the Cluster, without success. I have tried to configure it directly in the Spark Session, also without success. I also followed the instructions in the documentation to use a repository within Databricks, but since the base project does not use this Hook, it did not give the error. Has anyone had a similar error? I thought I could run if I turned the project into a Wheel, but I can't use the "kedro package" since the project can't run inside Databricks. I would be grateful for any ideas, thank you!Matheus Sampaio
12/19/2022, 2:53 PMkedro nodes execution order
on a Databricks Notebook?
Thanksuser
12/19/2022, 5:08 PMDhaval Thakkar
12/20/2022, 8:19 AMSimon Myway
12/20/2022, 8:51 AMFrits
12/20/2022, 9:24 AMDhaval Thakkar
12/20/2022, 10:26 AMhooks.py
file for great-expectations and register the hook in the settings.py
file
4. Execute the following command kedro run
5. Now I was expecting this to work directly but now I am getting this error `ValueError: Unable to load datasource files_datasource
-- no configuration found or invalid configuration.`
Please use the latest develop branch for the following project to look through the issue : https://github.com/DhavalThkkar/ecom-analytics
This is extremely difficult. Can someone guide me if I am doing anything wrong?Pedro Abreu
12/20/2022, 11:24 AMDhaval Thakkar
12/20/2022, 3:02 PMuser
12/20/2022, 3:38 PMDhaval Thakkar
12/21/2022, 10:08 AMDaniel Bull
12/21/2022, 12:00 PMSeth
12/21/2022, 1:11 PMAnu Arora
12/21/2022, 1:25 PMUsageError: Line magic function `%run_viz` not found.