You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Running test_group_apply_udf or test_group_apply_udf_more_types hangs in the Databricks 9.1 environment. There's no CPU utilization, so it is not an infinite loop. From a stack trace, it appears the code is waiting for data from Python that never arrives:
"Executor task launch worker for task 1.0 in stage 6.0 (TID 45)" #79 daemon prio=5 os_prio=0 tid=0x00007fab6c140000 nid=0x5876 runnable [0x00007faaa43be000]
java.lang.Thread.State: RUNNABLE
at java.net.SocketInputStream.socketRead0(Native Method)
at java.net.SocketInputStream.socketRead(SocketInputStream.java:116)
at java.net.SocketInputStream.read(SocketInputStream.java:171)
at java.net.SocketInputStream.read(SocketInputStream.java:141)
at java.io.BufferedInputStream.fill(BufferedInputStream.java:246)
at java.io.BufferedInputStream.read(BufferedInputStream.java:265)
- locked <0x00000000fa10ad40> (a java.io.BufferedInputStream)
at java.io.DataInputStream.readInt(DataInputStream.java:387)
at org.apache.spark.sql.rapids.execution.python.GpuPythonArrowOutput$$anon$1.read(GpuArrowEvalPythonExec.scala:328)
at org.apache.spark.sql.rapids.execution.python.GpuPythonArrowOutput$$anon$1.read(GpuArrowEvalPythonExec.scala:285)
The text was updated successfully, but these errors were encountered:
There is a DB specific config spark.databricks.execution.pandasZeroConfConversion.groupbyApply.enabled which is false by default. The test can pass after setting this config to true.
Seems DB 9.1 supports to disable this 'zero-conf-conversion' feature and has it disabled by default. While the plugin is missing the support of disabling it.
That is to say, the correct Python runner (grouped python runner or base arrow python runner) should be picked according to this config when being created in the GpuFlatMapGroupInPandas operator.
Running
test_group_apply_udf
ortest_group_apply_udf_more_types
hangs in the Databricks 9.1 environment. There's no CPU utilization, so it is not an infinite loop. From a stack trace, it appears the code is waiting for data from Python that never arrives:The text was updated successfully, but these errors were encountered: