| ID ▴ | Description | Submitted | Duration | Job IDs | Sub Execution IDs |
|---|---|---|---|---|---|
| 0 |
DataFrameCommonJob_checkSqlDbOncreation
DataFrameCommonJob_checkSqlDbOncreation |
2026/02/19 12:04:33 | 2 s | [0] |
|
| 2 |
DataFrameCommonJob_checkSqlDbOncreation
DataFrameCommonJob_checkSqlDbOncreation |
2026/02/19 12:05:45 | 1 s | [2] |
|
| 3 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:08:57 | 25 ms |
|
|
| 4 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:08:57 | 2 ms |
|
|
| 5 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:08:57 | 1 s | [3] |
|
| 6 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:08:57 | 2 s | [4][5] |
|
| 7 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:09:18 | 0 ms |
|
|
| 8 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:09:18 | 1 s | [6][8] |
|
| 9 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:19 | 1 ms |
|
|
| 10 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:19 | 1 s | [7] |
|
| 11 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:37 | 0 ms |
|
|
| 12 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:37 | 1 s | [9] |
|
| 13 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:09:59 | 0 ms |
|
|
| 14 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:09:59 | 1 s | [10][12] |
|
| 15 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:59 | 1 ms |
|
|
| 16 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:59 | 1 s | [11] |
|
| 17 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:11:12 | 0 ms |
|
|
| 18 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:11:12 | 1 s | [13][15] |
|
| 19 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:11:13 | 0 ms |
|
|
| 20 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:11:13 | 2 s | [14] |
|
| 21 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:11:33 | 0 ms |
|
|
| 22 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:11:33 | 1 s | [16][18] |
|
| 23 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:11:33 | 0 ms |
|
|
| 24 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:11:33 | 1.0 min | [17] |
|
| 25 |
PersistFrame
PersistFrame |
2026/02/19 12:13:29 | 60 s | [19] |
|
| 26 |
PersistFrame
PersistFrame |
2026/02/19 12:13:29 | 3 s | [20] |
|
| 27 |
PersistFrame
PersistFrame |
2026/02/19 12:13:30 | 1.1 min | [21] |
|
| 28 |
PersistFrame
PersistFrame |
2026/02/19 12:13:30 | 3 s | [22] |
|
| 29 |
PersistFrame
PersistFrame |
2026/02/19 12:13:31 | 2 s | [23] |
|
| 30 |
PersistFrame
PersistFrame |
2026/02/19 12:13:34 | 1 s | [24] |
|
| 31 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:14:44 | 0 ms |
|
|
| 32 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:14:44 | 1 ms |
|
|
| 33 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 12:14:44 | 0.4 s | [27][29] |
|
| 34 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:14:44 | 0.5 s | [28] |
|
| 35 |
PersistFrame
PersistFrame |
2026/02/19 12:23:19 | 1 s | [30] |
|
| 36 |
PersistFrame
PersistFrame |
2026/02/19 12:23:20 | 1 s | [31] |
|
| 37 |
PersistFrame
PersistFrame |
2026/02/19 12:23:20 | 1 s | [32] |
|
| 38 |
PersistFrame
PersistFrame |
2026/02/19 12:23:21 | 3.5 min | [33] |
|
| 39 |
PersistFrame
PersistFrame |
2026/02/19 12:23:21 | 4 s | [34] |
|
| 40 |
PersistFrame
PersistFrame |
2026/02/19 12:23:22 | 6.3 min | [35] |
|
| 41 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:16:34 | 1 ms |
|
|
| 42 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:16:34 | 0.1 s | [37][38] |
|
| 43 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:16:34 | 1 ms |
|
|
| 44 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:16:35 | 0.2 s | [40] |
|
| 45 |
DataFrameCommonJob_checkCSVSource
DataFrameCommonJob_checkCSVSource |
2026/02/19 13:19:47 | 0.2 s | [41] |
|
| 46 |
DataFrameCommonJob_checkCSVSource
DataFrameCommonJob_checkCSVSource |
2026/02/19 13:19:48 | 63 ms | [43] |
|
| 47 |
DataFrameCommonJob_checkCSVSource
DataFrameCommonJob_checkCSVSource |
2026/02/19 13:19:48 | 0.1 s | [44] |
|
| 48 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:20:25 | 0 ms |
|
|
| 49 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:20:25 | 0.1 s | [46][47] |
|
| 50 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:20:25 | 0 ms |
|
|
| 51 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:20:25 | 0.1 s | [49] |
|
| 52 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:20:46 | 0 ms |
|
|
| 53 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:20:46 | 88 ms | [51][52] |
|
| 54 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:20:47 | 0 ms |
|
|
| 55 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:20:47 | 0.1 s | [54] |
|
| 56 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:21:42 | 51 ms | [55] |
|
| 57 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:21:42 | 0 ms |
|
|
| 58 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:21:42 | 0.2 s | [56][58] |
|
| 59 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:21:42 | 46 ms | [57] |
|
| 60 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:21:43 | 0 ms |
|
|
| 61 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:21:43 | 0.1 s | [59] |
|
| 62 |
DataFrameCommonJob_dataExecValueAnalysis
DataFrameCommonJob_dataExecValueAnalysis |
2026/02/19 13:21:56 | 0 ms |
|
|
| 63 |
DataFrameCommonJob_dataExecValueAnalysis
DataFrameCommonJob_dataExecValueAnalysis |
2026/02/19 13:21:56 | 0.2 s | [61][62] |
|
| 64 |
DataFrameCommonJob_dataExecValueAnalysis
DataFrameCommonJob_dataExecValueAnalysis |
2026/02/19 13:21:58 | 0 ms |
|
|
| 65 |
DataFrameCommonJob_dataExecValueAnalysis
DataFrameCommonJob_dataExecValueAnalysis |
2026/02/19 13:21:58 | 0.5 s | [64][65] |
|
| 66 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:25:29 | 0 ms |
|
|
| 67 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:25:29 | 88 ms | [67][68] |
|
| 68 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:25:29 | 0 ms |
|
|
| 69 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:25:29 | 0.1 s | [70] |
|
| 70 |
DataFrameCommonJob_dataExecSummaryOnCreation
DataFrameCommonJob_dataExecSummaryOnCreation |
2026/02/19 13:26:28 | 1 ms |
|
|
| 71 |
DataFrameCommonJob_dataExecSummaryOnCreation
DataFrameCommonJob_dataExecSummaryOnCreation |
2026/02/19 13:26:28 | 79 ms | [72][73] |
|
| 72 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/19 13:26:29 | 1 ms |
|
|
| 73 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/19 13:26:29 | 68 ms | [75] |
|
| 74 |
DataFrameCommonJob_dataExecSummaryOnCreation
DataFrameCommonJob_dataExecSummaryOnCreation |
2026/02/19 13:29:06 | 0 ms |
|
|
| 75 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/19 13:29:06 | 1 ms |
|
|
| 76 |
DataFrameCommonJob_dataExecSummaryOnCreation
DataFrameCommonJob_dataExecSummaryOnCreation |
2026/02/19 13:29:06 | 94 ms | [78][79] |
|
| 77 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/19 13:29:06 | 0.2 s | [80] |
|
| 78 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:29:35 | 1 ms |
|
|
| 79 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:29:35 | 83 ms | [82][83] |
|
| 80 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:29:35 | 0 ms |
|
|
| 81 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:29:35 | 74 ms | [85] |
|
| 82 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:34:22 | 0 ms |
|
|
| 83 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:34:22 | 0.1 s | [87][88] |
|
| 84 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:34:23 | 1 ms |
|
|
| 85 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:34:23 | 0.8 s | [90][91] |
|
| 86 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:48:15 | 0 ms |
|
|
| 87 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 13:48:15 | 0.1 s | [93][95] |
|
| 88 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:48:15 | 0 ms |
|
|
| 89 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:48:15 | 63 ms | [96] |
|
| 90 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/19 13:50:55 | 1 ms |
|
|
| 91 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/19 13:50:55 | 0 ms |
|
|
| 92 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/19 13:50:55 | 0.2 s | [98] |
|
| 93 |
DataFrameCommonJob_dataExecSummaryOnCreation
DataFrameCommonJob_dataExecSummaryOnCreation |
2026/02/19 13:51:19 | 0 ms |
|
|
| 94 |
DataFrameCommonJob_dataExecSummaryOnCreation
DataFrameCommonJob_dataExecSummaryOnCreation |
2026/02/19 13:51:19 | 0 ms |
|
|
| 95 |
DataFrameCommonJob_dataExecSummaryOnCreation
DataFrameCommonJob_dataExecSummaryOnCreation |
2026/02/19 13:51:19 | 76 ms | [100][101] |
|
| 96 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 14:00:36 | 1 ms |
|
|
| 97 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/19 14:00:36 | 78 ms | [103][104] |
|
| 98 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 14:00:36 | 1 ms |
|
|
| 99 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 14:00:36 | 0.2 s | [106] |
|
| 100 |
DataFrameExportJob_exec_file
DataFrameExportJob_exec_file |
2026/02/19 14:11:55 | 61 ms | [107] |
|
| ID ▾ | Description | Submitted | Duration | Succeeded Job IDs | Failed Job IDs | Error Message | Sub Execution IDs |
|---|---|---|---|---|---|---|---|
| 659 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:46:52 | 26 s | [801] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 657 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:46:52 | 27 s | [800] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.$anonfun$count$1(Dataset.scala:3616) at org.apache.spark.sql.Dataset.$anonfun$count$1$adapted(Dataset.scala:3615) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.count(Dataset.scala:3615) at com.adb.hdh.spark.job.impl.ColumnStatJob.stats(ColumnStatJob.java:47) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:35) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 607 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:24:36 | 20 s | [727] | [728] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
| 605 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:24:36 | 22 s | [726] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.$anonfun$count$1(Dataset.scala:3616) at org.apache.spark.sql.Dataset.$anonfun$count$1$adapted(Dataset.scala:3615) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.count(Dataset.scala:3615) at com.adb.hdh.spark.job.impl.ColumnStatJob.stats(ColumnStatJob.java:47) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:35) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 571 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:18:00 | 43 s | [674] | [675] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
| 569 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:18:00 | 44 s | [672] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.$anonfun$count$1(Dataset.scala:3616) at org.apache.spark.sql.Dataset.$anonfun$count$1$adapted(Dataset.scala:3615) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.count(Dataset.scala:3615) at com.adb.hdh.spark.job.impl.ColumnStatJob.stats(ColumnStatJob.java:47) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:35) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 466 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 11:26:01 | 38 s | [548] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 445 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 11:18:32 | 32 s | [530] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.checkSqlDb(DataFrameCommonJob.java:276) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:183) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 400 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 10:33:52 | 1 s | [488] | Job aborted due to stage failure: Task 0 in stage 604.0 failed 1 times, most recent failure: Lost task 0.0 in stage 604.0 (TID 607) (daedd42205fa executor driver): java.sql.SQLSyntaxErrorException: ORA-00923: FROM keyword not found where expectedJob aborted due to stage failure: Task 0 in stage 604.0 failed 1 times, most recent failure: Lost task 0.0 in stage 604.0 (TID 607) (daedd42205fa executor driver): java.sql.SQLSyntaxErrorException: ORA-00923: FROM keyword not found where expected
at oracle.jdbc.driver.T4CTTIoer11.processError(T4CTTIoer11.java:526)
at oracle.jdbc.driver.T4CTTIoer11.processError(T4CTTIoer11.java:462)
at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:1104)
at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:553)
at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:269)
at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:655)
at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:270)
at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:91)
at oracle.jdbc.driver.T4CPreparedStatement.executeForDescribe(T4CPreparedStatement.java:807)
at oracle.jdbc.driver.OracleStatement.executeMaybeDescribe(OracleStatement.java:991)
at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1176)
at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3671)
at oracle.jdbc.driver.T4CPreparedStatement.executeInternal(T4CPreparedStatement.java:1426)
at oracle.jdbc.driver.OraclePreparedStatement.executeQuery(OraclePreparedStatement.java:3718)
at oracle.jdbc.driver.OraclePreparedStatementWrapper.executeQuery(OraclePreparedStatementWrapper.java:1167)
at org.apache.spark.sql.execution.datasources.jdbc.JDBCRDD.compute(JDBCRDD.scala:304)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:93)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:621)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:624)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.base/java.lang.Thread.run(Unknown Source)
Caused by: Error : 923, Position : 26, Sql = SELECT "RAWTOHEX("VISIT_ID")","PATIENT_ID","IN_DATE","OUT_DATE","IN_MODE","OUT_MODE","IN_UNIT","OUT_UNIT","VISIT_TYPE","CHAMP_PMSI","UNIT_PATH","EMERGENCY","CANCELED" FROM (SELECT RAWTOHEX("VISIT_ID"), "PATIENT_ID", "IN_DATE", "OUT_DATE", "IN_MODE", "OUT_MODE", "IN_UNIT", "OUT_UNIT", "VISIT_TYPE", "CHAMP_PMSI", "UNIT_PATH", "EMERGENCY", "CANCELED" FROM "DTM_622"."VISIT") x , OriginalSql = SELECT "RAWTOHEX("VISIT_ID")","PATIENT_ID","IN_DATE","OUT_DATE","IN_MODE","OUT_MODE","IN_UNIT","OUT_UNIT","VISIT_TYPE","CHAMP_PMSI","UNIT_PATH","EMERGENCY","CANCELED" FROM (SELECT RAWTOHEX("VISIT_ID"), "PATIENT_ID", "IN_DATE", "OUT_DATE", "IN_MODE", "OUT_MODE", "IN_UNIT", "OUT_UNIT", "VISIT_TYPE", "CHAMP_PMSI", "UNIT_PATH", "EMERGENCY", "CANCELED" FROM "DTM_622"."VISIT") x , Error Msg = ORA-00923: FROM keyword not found where expected
at oracle.jdbc.driver.T4CTTIoer11.processError(T4CTTIoer11.java:530)
... 34 more
Driver stacktrace:
|
|
|
| 1 |
DataFrameCommonJob_checkSqlDbOncreation
DataFrameCommonJob_checkSqlDbOncreation |
2026/02/19 12:04:47 | 50 s | [1] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.checkSqlDb(DataFrameCommonJob.java:276) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:183) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|