Skip to content

Error on PySpark deduping example: no jars directory in splink #864

Answered by mamonu
mmagoffin-sd asked this question in Q&A
Discussion options

You must be logged in to vote

Can you either

  • put the jar in the path :
    d:\Users\mmagoffin\AppData\Local\hatch\env\virtual\ds-misc-mmagoffin-S1xDGiIH\ds-misc-mmagoffin\lib\site-packages\splink\jars/

  • or change the path (where currently you store the output of the function similarity_jar_location)
    to your working directory.

Also a warning: when you define settings for the spark session you need to restart your kernel in
order for these to be updated. If you just change the path and run the cell you dont get the jars loaded into the session

Replies: 1 comment 6 replies

Comment options

You must be logged in to vote
6 replies
@RobinL
Comment options

@mmagoffin-sd
Comment options

@RobinL
Comment options

@mmagoffin-sd
Comment options

@mmagoffin-sd
Comment options

Answer selected by mmagoffin-sd
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
3 participants