Skip to content
Advertisement

Interpolation in PySpark throws java.lang.IllegalArgumentException

I don’t know how to interpolate in PySpark when the DataFrame contains many columns. Let me xplain.

JavaScript

I need to group by webID and interpolate counts values at 1 minute interval. However, when I apply the below-shown code,

JavaScript

Error:

JavaScript

Advertisement

Answer

Set the environment variable ARROW_PRE_0_15_IPC_FORMAT=1.

https://spark.apache.org/docs/3.0.0-preview/sql-pyspark-pandas-with-arrow.html#compatibiliy-setting-for-pyarrow–0150-and-spark-23x-24x

JavaScript
User contributions licensed under: CC BY-SA
1 People found this is helpful
Advertisement