I have a project that mainly uses Java. We implemented some Python functions in a package and would like to unit test them. There is a package called src/python where these .py files are located. I have to problems implementing the tests: How can I make sure that these unit tests integrate with the maven test framework? So maven will
Tag: maven
PySpark 2.x: Programmatically adding Maven JAR Coordinates to Spark
The following is my PySpark startup snippet, which is pretty reliable (I’ve been using it a long time). Today I added the two Maven Coordinates shown in the spark.jars.packages option (effectively “plugging” in Kafka support). Now that normally triggers dependency downloads (performed by Spark automatically): However the plugins aren’t downloading and/or loading when I run the snippet (e.g. ./python -i