| ID | Description ▴ | Submitted | Duration | Job IDs | Sub Execution IDs |
|---|---|---|---|---|---|
| 45 |
DataFrameCommonJob_checkCSVSource
DataFrameCommonJob_checkCSVSource |
2026/02/19 13:19:47 | 0.2 s | [41] |
|
| 46 |
DataFrameCommonJob_checkCSVSource
DataFrameCommonJob_checkCSVSource |
2026/02/19 13:19:48 | 63 ms | [43] |
|
| 47 |
DataFrameCommonJob_checkCSVSource
DataFrameCommonJob_checkCSVSource |
2026/02/19 13:19:48 | 0.1 s | [44] |
|
| 325 |
DataFrameCommonJob_checkExtraction
DataFrameCommonJob_checkExtraction |
2026/02/20 08:21:22 | 88 ms | [406] |
|
| 345 |
DataFrameCommonJob_checkExtraction
DataFrameCommonJob_checkExtraction |
2026/02/20 10:02:20 | 54 ms | [428] |
|
| 320 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 08:03:10 | 1 s | [400] |
|
| 321 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 08:04:20 | 1 s | [401] |
|
| 322 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 08:18:00 | 1 s | [402] |
|
| 323 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 08:20:02 | 1 s | [403] |
|
| 326 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 09:33:03 | 1 s | [407] |
|
| 350 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 10:06:36 | 1 s | [434] |
|
| 373 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 10:18:38 | 2 s | [461] |
|
| 374 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 10:20:16 | 2 s | [462] |
|
| 401 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 10:34:19 | 1 s | [489] |
|
| 410 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 10:40:13 | 1 s | [498] |
|
| 415 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 10:40:46 | 1 s | [502] |
|
| 426 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 11:04:00 | 1 s | [514] |
|
| 427 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 11:10:12 | 1 s | [515] |
|
| 432 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 11:11:45 | 1 s | [519] |
|
| 433 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 11:12:14 | 1 s | [520] |
|
| 434 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 11:13:29 | 1 s | [521] |
|
| 452 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 11:23:24 | 2 s | [538] |
|
| 0 |
DataFrameCommonJob_checkSqlDbOncreation
DataFrameCommonJob_checkSqlDbOncreation |
2026/02/19 12:04:33 | 2 s | [0] |
|
| 2 |
DataFrameCommonJob_checkSqlDbOncreation
DataFrameCommonJob_checkSqlDbOncreation |
2026/02/19 12:05:45 | 1 s | [2] |
|
| 351 |
DataFrameCommonJob_checkSqlDbOncreation
DataFrameCommonJob_checkSqlDbOncreation |
2026/02/20 10:08:40 | 1 s | [435] |
|
| 352 |
DataFrameCommonJob_checkSqlDbOncreation
DataFrameCommonJob_checkSqlDbOncreation |
2026/02/20 10:09:47 | 1 s | [436] |
|
| 666 |
DataFrameCommonJob_checkSqlDbOncreation
DataFrameCommonJob_checkSqlDbOncreation |
2026/02/20 13:16:08 | 1 s | [812] |
|
| 4 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:08:57 | 2 ms |
|
|
| 5 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:08:57 | 1 s | [3] |
|
| 9 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:19 | 1 ms |
|
|
| 10 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:19 | 1 s | [7] |
|
| 11 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:37 | 0 ms |
|
|
| 12 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:37 | 1 s | [9] |
|
| 15 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:59 | 1 ms |
|
|
| 16 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:09:59 | 1 s | [11] |
|
| 19 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:11:13 | 0 ms |
|
|
| 20 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:11:13 | 2 s | [14] |
|
| 23 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:11:33 | 0 ms |
|
|
| 24 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:11:33 | 1.0 min | [17] |
|
| 32 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:14:44 | 1 ms |
|
|
| 34 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 12:14:44 | 0.5 s | [28] |
|
| 43 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:16:34 | 1 ms |
|
|
| 44 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:16:35 | 0.2 s | [40] |
|
| 50 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:20:25 | 0 ms |
|
|
| 51 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:20:25 | 0.1 s | [49] |
|
| 54 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:20:47 | 0 ms |
|
|
| 55 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:20:47 | 0.1 s | [54] |
|
| 59 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:21:42 | 46 ms | [57] |
|
| 60 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:21:43 | 0 ms |
|
|
| 61 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:21:43 | 0.1 s | [59] |
|
| 68 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:25:29 | 0 ms |
|
|
| 69 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:25:29 | 0.1 s | [70] |
|
| 80 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:29:35 | 0 ms |
|
|
| 81 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:29:35 | 74 ms | [85] |
|
| 84 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:34:23 | 1 ms |
|
|
| 85 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:34:23 | 0.8 s | [90][91] |
|
| 88 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:48:15 | 0 ms |
|
|
| 89 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 13:48:15 | 63 ms | [96] |
|
| 98 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 14:00:36 | 1 ms |
|
|
| 99 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 14:00:36 | 0.2 s | [106] |
|
| 107 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 14:14:24 | 0 ms |
|
|
| 108 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 14:14:24 | 0.1 s | [115] |
|
| 198 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:11:58 | 0 ms |
|
|
| 199 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:11:58 | 1 s | [218][219] |
|
| 202 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:16:19 | 0 ms |
|
|
| 203 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:16:19 | 2 s | [224][225] |
|
| 206 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:18:44 | 0 ms |
|
|
| 207 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:18:44 | 5 s | [230][231] |
|
| 210 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:19:54 | 0 ms |
|
|
| 211 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:19:54 | 0.6 s | [235][236] |
|
| 214 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:20:55 | 0 ms |
|
|
| 215 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:20:55 | 0.6 s | [241][242] |
|
| 218 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:21:17 | 0 ms |
|
|
| 219 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:21:17 | 0.5 s | [247][248] |
|
| 222 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:22:08 | 0 ms |
|
|
| 223 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:22:08 | 0.1 s | [254][255] |
|
| 226 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:25:29 | 0 ms |
|
|
| 227 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:25:29 | 6 s | [260][261] |
|
| 230 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:25:54 | 1 ms |
|
|
| 231 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:25:54 | 6 s | [265][266] |
|
| 234 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:31:14 | 1 ms |
|
|
| 235 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:31:14 | 6 s | [272][273] |
|
| 238 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:32:15 | 0 ms |
|
|
| 239 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:32:15 | 6 s | [277][278] |
|
| 242 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:36:56 | 0 ms |
|
|
| 243 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:36:56 | 7 s | [284][285] |
|
| 246 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:37:54 | 1 ms |
|
|
| 247 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:37:54 | 4 s | [290][291] |
|
| 250 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:39:33 | 0 ms |
|
|
| 251 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:39:33 | 4 s | [296][297] |
|
| 254 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:41:06 | 1 ms |
|
|
| 255 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:41:06 | 0.5 s | [301][302] |
|
| 258 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:42:12 | 0 ms |
|
|
| 259 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:42:13 | 0.6 s | [307][308] |
|
| 262 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:44:13 | 0 ms |
|
|
| 263 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:44:13 | 0.5 s | [313][314] |
|
| 266 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:44:41 | 0 ms |
|
|
| 267 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:44:41 | 0.6 s | [319][320] |
|
| 270 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:48:11 | 0 ms |
|
|
| 271 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/19 16:48:11 | 0.6 s | [325][326] |
|
| ID ▾ | Description | Submitted | Duration | Succeeded Job IDs | Failed Job IDs | Error Message | Sub Execution IDs |
|---|---|---|---|---|---|---|---|
| 659 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:46:52 | 26 s | [801] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 657 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:46:52 | 27 s | [800] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.$anonfun$count$1(Dataset.scala:3616) at org.apache.spark.sql.Dataset.$anonfun$count$1$adapted(Dataset.scala:3615) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.count(Dataset.scala:3615) at com.adb.hdh.spark.job.impl.ColumnStatJob.stats(ColumnStatJob.java:47) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:35) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 607 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:24:36 | 20 s | [727] | [728] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
| 605 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:24:36 | 22 s | [726] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.$anonfun$count$1(Dataset.scala:3616) at org.apache.spark.sql.Dataset.$anonfun$count$1$adapted(Dataset.scala:3615) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.count(Dataset.scala:3615) at com.adb.hdh.spark.job.impl.ColumnStatJob.stats(ColumnStatJob.java:47) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:35) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 571 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:18:00 | 43 s | [674] | [675] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
| 569 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:18:00 | 44 s | [672] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.$anonfun$count$1(Dataset.scala:3616) at org.apache.spark.sql.Dataset.$anonfun$count$1$adapted(Dataset.scala:3615) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.count(Dataset.scala:3615) at com.adb.hdh.spark.job.impl.ColumnStatJob.stats(ColumnStatJob.java:47) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:35) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 466 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 11:26:01 | 38 s | [548] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 445 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 11:18:32 | 32 s | [530] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.checkSqlDb(DataFrameCommonJob.java:276) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:183) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 400 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 10:33:52 | 1 s | [488] | Job aborted due to stage failure: Task 0 in stage 604.0 failed 1 times, most recent failure: Lost task 0.0 in stage 604.0 (TID 607) (daedd42205fa executor driver): java.sql.SQLSyntaxErrorException: ORA-00923: FROM keyword not found where expectedJob aborted due to stage failure: Task 0 in stage 604.0 failed 1 times, most recent failure: Lost task 0.0 in stage 604.0 (TID 607) (daedd42205fa executor driver): java.sql.SQLSyntaxErrorException: ORA-00923: FROM keyword not found where expected
at oracle.jdbc.driver.T4CTTIoer11.processError(T4CTTIoer11.java:526)
at oracle.jdbc.driver.T4CTTIoer11.processError(T4CTTIoer11.java:462)
at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:1104)
at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:553)
at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:269)
at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:655)
at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:270)
at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:91)
at oracle.jdbc.driver.T4CPreparedStatement.executeForDescribe(T4CPreparedStatement.java:807)
at oracle.jdbc.driver.OracleStatement.executeMaybeDescribe(OracleStatement.java:991)
at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1176)
at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3671)
at oracle.jdbc.driver.T4CPreparedStatement.executeInternal(T4CPreparedStatement.java:1426)
at oracle.jdbc.driver.OraclePreparedStatement.executeQuery(OraclePreparedStatement.java:3718)
at oracle.jdbc.driver.OraclePreparedStatementWrapper.executeQuery(OraclePreparedStatementWrapper.java:1167)
at org.apache.spark.sql.execution.datasources.jdbc.JDBCRDD.compute(JDBCRDD.scala:304)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:93)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:621)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:624)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.base/java.lang.Thread.run(Unknown Source)
Caused by: Error : 923, Position : 26, Sql = SELECT "RAWTOHEX("VISIT_ID")","PATIENT_ID","IN_DATE","OUT_DATE","IN_MODE","OUT_MODE","IN_UNIT","OUT_UNIT","VISIT_TYPE","CHAMP_PMSI","UNIT_PATH","EMERGENCY","CANCELED" FROM (SELECT RAWTOHEX("VISIT_ID"), "PATIENT_ID", "IN_DATE", "OUT_DATE", "IN_MODE", "OUT_MODE", "IN_UNIT", "OUT_UNIT", "VISIT_TYPE", "CHAMP_PMSI", "UNIT_PATH", "EMERGENCY", "CANCELED" FROM "DTM_622"."VISIT") x , OriginalSql = SELECT "RAWTOHEX("VISIT_ID")","PATIENT_ID","IN_DATE","OUT_DATE","IN_MODE","OUT_MODE","IN_UNIT","OUT_UNIT","VISIT_TYPE","CHAMP_PMSI","UNIT_PATH","EMERGENCY","CANCELED" FROM (SELECT RAWTOHEX("VISIT_ID"), "PATIENT_ID", "IN_DATE", "OUT_DATE", "IN_MODE", "OUT_MODE", "IN_UNIT", "OUT_UNIT", "VISIT_TYPE", "CHAMP_PMSI", "UNIT_PATH", "EMERGENCY", "CANCELED" FROM "DTM_622"."VISIT") x , Error Msg = ORA-00923: FROM keyword not found where expected
at oracle.jdbc.driver.T4CTTIoer11.processError(T4CTTIoer11.java:530)
... 34 more
Driver stacktrace:
|
|
|
| 1 |
DataFrameCommonJob_checkSqlDbOncreation
DataFrameCommonJob_checkSqlDbOncreation |
2026/02/19 12:04:47 | 50 s | [1] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.checkSqlDb(DataFrameCommonJob.java:276) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:183) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|