WARNING:splink.internals.spark.database_api:Unable to load custom Spark SQL functions such as jaro_winkler #2433
Unanswered
MariaWerner-Wien
asked this question in
Q&A
Replies: 1 comment 2 replies
-
Hi, |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello,
I am trying to run Splink on PySpark but every time I get one and the same error. I tried to implement the example in the tutorial provided in the Splink documentation both in version 4 and version 3. In particular, I tried:
I tested it in:
I get the following error:
The GitHub link provided in the error message doesn't work.
It seems the
similarity_jar_location
is not located inspark.jar_location
. Instead, I found it ininternals
:from splink.internals.spark.jar_location import similarity_jar_location
.Nonetheless, the error persists.
Any suggestions how to overcome this issue would be highly appreciated. I need to run Splink on a Spark cluster.
Many thanks in advance!
Beta Was this translation helpful? Give feedback.
All reactions