| ID | Description | Submitted | Duration | Job IDs ▾ | Sub Execution IDs |
|---|---|---|---|---|---|
| 756 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 13:46:38 | 1 s | [927] |
|
| 754 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:46:37 | 1 s | [926][928] |
|
| 752 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/20 13:46:14 | 1 s | [924] |
|
| 750 |
DataFrameCommonJob_dataExecSummaryOnCreation
DataFrameCommonJob_dataExecSummaryOnCreation |
2026/02/20 13:46:14 | 1 s | [923][925] |
|
| 748 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/20 13:41:20 | 0.2 s | [920][921][922] |
|
| 745 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/20 13:41:20 | 0.2 s | [917][918][919] |
|
| 741 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/20 13:41:20 | 0.2 s | [913][914][916] |
|
| 738 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/20 13:41:19 | 0.2 s | [909][910][912] |
|
| 732 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/20 13:40:58 | 1 s | [901][902] |
|
| 729 |
DataFrameCommonJob_dataExecColumnAnalysis
DataFrameCommonJob_dataExecColumnAnalysis |
2026/02/20 13:35:44 | 1 s | [897][899][900] |
|
| 727 |
DataFrameCommonJob_dataExecCountALL
DataFrameCommonJob_dataExecCountALL |
2026/02/20 13:35:44 | 1 s | [896][898] |
|
| 725 |
DataFrameCommonJob_dataExecColumnAnalysis
DataFrameCommonJob_dataExecColumnAnalysis |
2026/02/20 13:35:43 | 0.4 s | [893][894][895] |
|
| 724 |
DataFrameCommonJob_dataExecCountALL
DataFrameCommonJob_dataExecCountALL |
2026/02/20 13:35:43 | 0.2 s | [891][892] |
|
| 721 |
DataFrameCommonJob_dataExecColumnAnalysis
DataFrameCommonJob_dataExecColumnAnalysis |
2026/02/20 13:34:56 | 1 s | [887][888] |
|
| 719 |
DataFrameCommonJob_dataExecColumnAnalysis
DataFrameCommonJob_dataExecColumnAnalysis |
2026/02/20 13:34:54 | 1 s | [884][886] |
|
| 717 |
DataFrameCommonJob_dataExecCountALL
DataFrameCommonJob_dataExecCountALL |
2026/02/20 13:34:53 | 1 s | [883][885] |
|
| 715 |
DataFrameCommonJob_dataExecCountALL
DataFrameCommonJob_dataExecCountALL |
2026/02/20 13:34:51 | 1 s | [881][882] |
|
| 713 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 13:33:50 | 0.4 s | [877][878][880] |
|
| 710 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:33:50 | 0.3 s | [873][874][876][879] |
|
| 707 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 13:32:40 | 0.3 s | [868][869] |
|
| 705 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:32:40 | 89 ms | [865][866] |
|
| 703 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 13:32:02 | 84 ms | [863] |
|
| 701 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:32:02 | 56 ms | [860][861] |
|
| 699 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 13:29:14 | 1 s | [857] |
|
| 697 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:29:14 | 1 s | [856][858] |
|
| 695 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 13:27:38 | 94 ms | [855] |
|
| 693 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:27:38 | 63 ms | [852][853] |
|
| 691 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:27:15 | 0.3 s | [849][850] |
|
| 690 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:27:14 | 55 ms | [847][848] |
|
| 689 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:27:14 | 66 ms | [845][846] |
|
| 688 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:27:14 | 54 ms | [843][844] |
|
| 687 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:27:14 | 75 ms | [841][842] |
|
| 686 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:27:14 | 77 ms | [839][840] |
|
| 685 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:27:14 | 0.2 s | [837][838] |
|
| 684 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 13:27:14 | 73 ms | [836] |
|
| 682 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:27:14 | 0.2 s | [834][835] |
|
| 681 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:27:14 | 76 ms | [831][832] |
|
| 679 |
DataFrameCommonJob_dataExecValueAnalysis
DataFrameCommonJob_dataExecValueAnalysis |
2026/02/20 13:26:46 | 0.3 s | [827][828][829] |
|
| 677 |
DataFrameCommonJob_dataExecValueAnalysis
DataFrameCommonJob_dataExecValueAnalysis |
2026/02/20 13:26:42 | 0.3 s | [824][825] |
|
| 675 |
DataFrameCommonJob_dataExecValueAnalysis
DataFrameCommonJob_dataExecValueAnalysis |
2026/02/20 13:26:41 | 0.2 s | [821][822] |
|
| 673 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 13:21:39 | 64 ms | [819] |
|
| 671 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 13:21:39 | 66 ms | [816][817] |
|
| 669 |
PersistFrame
PersistFrame |
2026/02/20 13:21:21 | 1 s | [814] |
|
| 668 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 13:20:45 | 1 s | [813] |
|
| 666 |
DataFrameCommonJob_checkSqlDbOncreation
DataFrameCommonJob_checkSqlDbOncreation |
2026/02/20 13:16:08 | 1 s | [812] |
|
| 665 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/20 12:48:23 | 0.1 s | [810][811] |
|
| 663 |
DataFrameCommonJob_dataExecSummaryOnCreation
DataFrameCommonJob_dataExecSummaryOnCreation |
2026/02/20 12:48:23 | 77 ms | [807][808] |
|
| 661 |
PersistFrame
PersistFrame |
2026/02/20 12:47:54 | 23 s | [804][805] |
|
| 655 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:44:30 | 25 s | [795][797] |
|
| 653 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:44:30 | 24 s | [794][796] |
|
| 651 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:41:07 | 26 s | [789][790] |
|
| 649 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:41:07 | 24 s | [787][791] |
|
| 645 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:40:54 | 44 ms | [784] |
|
| 643 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:40:11 | 25 s | [780][782] |
|
| 641 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:40:11 | 24 s | [779][781] |
|
| 639 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:37:10 | 24 s | [774][776] |
|
| 637 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:37:10 | 24 s | [773][775] |
|
| 635 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:35:53 | 24 s | [768][770] |
|
| 633 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:35:52 | 24 s | [767][769] |
|
| 631 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:34:44 | 25 s | [762][763] |
|
| 629 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:34:44 | 26 s | [761][764] |
|
| 627 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:32:37 | 27 s | [756][757] |
|
| 625 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:32:37 | 25 s | [754][758] |
|
| 623 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:31:22 | 26 s | [750][751] |
|
| 621 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:31:22 | 24 s | [748][752] |
|
| 619 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:30:14 | 27 s | [744][745] |
|
| 617 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:30:13 | 27 s | [742][746] |
|
| 615 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:28:14 | 24 s | [738][739] |
|
| 613 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:28:14 | 24 s | [736][740] |
|
| 611 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:25:05 | 27 s | [732][733] |
|
| 609 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:25:05 | 24 s | [730][734] |
|
| 603 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:23:26 | 0.5 s | [721][723] |
|
| 601 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:23:26 | 0.4 s | [719][722] |
|
| 599 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:22:16 | 0.6 s | [715][717] |
|
| 597 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:22:15 | 0.4 s | [713][716] |
|
| 595 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:20:39 | 0.5 s | [709][711] |
|
| 593 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:20:38 | 0.3 s | [707][710] |
|
| 591 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:20:29 | 0.3 s | [703][704] |
|
| 589 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:20:29 | 0.3 s | [701][705] |
|
| 587 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:19:51 | 0.4 s | [697][698] |
|
| 585 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:19:51 | 0.4 s | [695][699] |
|
| 583 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:19:18 | 0.4 s | [691][692] |
|
| 581 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:19:18 | 0.4 s | [689][693] |
|
| 579 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:19:15 | 0.4 s | [685][686] |
|
| 577 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:19:14 | 0.4 s | [683][687] |
|
| 575 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:18:47 | 93 ms | [680][681] |
|
| 573 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:18:47 | 0.1 s | [677][679] |
|
| 567 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:16:05 | 27 s | [669][670] |
|
| 565 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:16:05 | 70 ms | [666][667] |
|
| 563 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:14:24 | 1 s | [663][664] |
|
| 561 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:14:24 | 75 ms | [660][661] |
|
| 559 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:13:43 | 0.1 s | [658] |
|
| 557 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:13:43 | 0.2 s | [656][657] |
|
| 555 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/20 12:11:24 | 1 s | [650][651][653] |
|
| 553 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/20 12:11:24 | 1 s | [648][649][652] |
|
| 546 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/20 12:09:29 | 1 s | [641][642][644] |
|
| 547 |
DataFrameCommonJob_dataExecSummaryOnCreation
DataFrameCommonJob_dataExecSummaryOnCreation |
2026/02/20 12:09:29 | 1 s | [639][640][643][645] |
|
| 539 |
DataFrameCommonJob_dataExecOnCreation
DataFrameCommonJob_dataExecOnCreation |
2026/02/20 12:08:15 | 1 s | [632][633][635] |
|
| 538 |
DataFrameCommonJob_dataExecSummaryOnCreation
DataFrameCommonJob_dataExecSummaryOnCreation |
2026/02/20 12:08:15 | 1 s | [630][631][634][636] |
|
| 531 |
DataFrameCommonJob_dataExecValueAnalysis
DataFrameCommonJob_dataExecValueAnalysis |
2026/02/20 12:07:48 | 2 s | [624][625][626][627] |
|
| ID ▾ | Description | Submitted | Duration | Succeeded Job IDs | Failed Job IDs | Error Message | Sub Execution IDs |
|---|---|---|---|---|---|---|---|
| 659 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:46:52 | 26 s | [801] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 657 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:46:52 | 27 s | [800] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.$anonfun$count$1(Dataset.scala:3616) at org.apache.spark.sql.Dataset.$anonfun$count$1$adapted(Dataset.scala:3615) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.count(Dataset.scala:3615) at com.adb.hdh.spark.job.impl.ColumnStatJob.stats(ColumnStatJob.java:47) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:35) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 607 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:24:36 | 20 s | [727] | [728] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
| 605 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:24:36 | 22 s | [726] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.$anonfun$count$1(Dataset.scala:3616) at org.apache.spark.sql.Dataset.$anonfun$count$1$adapted(Dataset.scala:3615) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.count(Dataset.scala:3615) at com.adb.hdh.spark.job.impl.ColumnStatJob.stats(ColumnStatJob.java:47) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:35) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 571 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 12:18:00 | 43 s | [674] | [675] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
| 569 |
DataFrameCommonJob_dataExecSummary
DataFrameCommonJob_dataExecSummary |
2026/02/20 12:18:00 | 44 s | [672] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(Unknown Source) at java.base/java.util.concurrent.LinkedBlockingQueue.take(Unknown Source) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.$anonfun$getFinalPhysicalPlan$1(AdaptiveSparkPlanExec.scala:322) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.getFinalPhysicalPlan(AdaptiveSparkPlanExec.scala:272) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.withFinalPlanUpdate(AdaptiveSparkPlanExec.scala:419) at org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanExec.executeCollect(AdaptiveSparkPlanExec.scala:392) at org.apache.spark.sql.Dataset.$anonfun$count$1(Dataset.scala:3616) at org.apache.spark.sql.Dataset.$anonfun$count$1$adapted(Dataset.scala:3615) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.count(Dataset.scala:3615) at com.adb.hdh.spark.job.impl.ColumnStatJob.stats(ColumnStatJob.java:47) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:35) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 466 |
DataFrameCommonJob_dataExec
DataFrameCommonJob_dataExec |
2026/02/20 11:26:01 | 38 s | [548] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.ExecDataJob.exec(ExecDataJob.java:39) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:142) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 445 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 11:18:32 | 32 s | [530] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.checkSqlDb(DataFrameCommonJob.java:276) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:183) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|
|
| 400 |
DataFrameCommonJob_checkSqlDb
DataFrameCommonJob_checkSqlDb |
2026/02/20 10:33:52 | 1 s | [488] | Job aborted due to stage failure: Task 0 in stage 604.0 failed 1 times, most recent failure: Lost task 0.0 in stage 604.0 (TID 607) (daedd42205fa executor driver): java.sql.SQLSyntaxErrorException: ORA-00923: FROM keyword not found where expectedJob aborted due to stage failure: Task 0 in stage 604.0 failed 1 times, most recent failure: Lost task 0.0 in stage 604.0 (TID 607) (daedd42205fa executor driver): java.sql.SQLSyntaxErrorException: ORA-00923: FROM keyword not found where expected
at oracle.jdbc.driver.T4CTTIoer11.processError(T4CTTIoer11.java:526)
at oracle.jdbc.driver.T4CTTIoer11.processError(T4CTTIoer11.java:462)
at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:1104)
at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:553)
at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:269)
at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:655)
at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:270)
at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:91)
at oracle.jdbc.driver.T4CPreparedStatement.executeForDescribe(T4CPreparedStatement.java:807)
at oracle.jdbc.driver.OracleStatement.executeMaybeDescribe(OracleStatement.java:991)
at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1176)
at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3671)
at oracle.jdbc.driver.T4CPreparedStatement.executeInternal(T4CPreparedStatement.java:1426)
at oracle.jdbc.driver.OraclePreparedStatement.executeQuery(OraclePreparedStatement.java:3718)
at oracle.jdbc.driver.OraclePreparedStatementWrapper.executeQuery(OraclePreparedStatementWrapper.java:1167)
at org.apache.spark.sql.execution.datasources.jdbc.JDBCRDD.compute(JDBCRDD.scala:304)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:93)
at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
at org.apache.spark.scheduler.Task.run(Task.scala:141)
at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:621)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:624)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.base/java.lang.Thread.run(Unknown Source)
Caused by: Error : 923, Position : 26, Sql = SELECT "RAWTOHEX("VISIT_ID")","PATIENT_ID","IN_DATE","OUT_DATE","IN_MODE","OUT_MODE","IN_UNIT","OUT_UNIT","VISIT_TYPE","CHAMP_PMSI","UNIT_PATH","EMERGENCY","CANCELED" FROM (SELECT RAWTOHEX("VISIT_ID"), "PATIENT_ID", "IN_DATE", "OUT_DATE", "IN_MODE", "OUT_MODE", "IN_UNIT", "OUT_UNIT", "VISIT_TYPE", "CHAMP_PMSI", "UNIT_PATH", "EMERGENCY", "CANCELED" FROM "DTM_622"."VISIT") x , OriginalSql = SELECT "RAWTOHEX("VISIT_ID")","PATIENT_ID","IN_DATE","OUT_DATE","IN_MODE","OUT_MODE","IN_UNIT","OUT_UNIT","VISIT_TYPE","CHAMP_PMSI","UNIT_PATH","EMERGENCY","CANCELED" FROM (SELECT RAWTOHEX("VISIT_ID"), "PATIENT_ID", "IN_DATE", "OUT_DATE", "IN_MODE", "OUT_MODE", "IN_UNIT", "OUT_UNIT", "VISIT_TYPE", "CHAMP_PMSI", "UNIT_PATH", "EMERGENCY", "CANCELED" FROM "DTM_622"."VISIT") x , Error Msg = ORA-00923: FROM keyword not found where expected
at oracle.jdbc.driver.T4CTTIoer11.processError(T4CTTIoer11.java:530)
... 34 more
Driver stacktrace:
|
|
|
| 1 |
DataFrameCommonJob_checkSqlDbOncreation
DataFrameCommonJob_checkSqlDbOncreation |
2026/02/19 12:04:47 | 50 s | [1] | java.lang.InterruptedExceptionjava.lang.InterruptedException at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(Unknown Source) at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(Unknown Source) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:242) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:258) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:187) at org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:342) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:986) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2393) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2414) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2433) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:530) at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:483) at org.apache.spark.sql.execution.CollectLimitExec.executeCollect(limit.scala:61) at org.apache.spark.sql.Dataset.collectFromPlan(Dataset.scala:4333) at org.apache.spark.sql.Dataset.$anonfun$head$1(Dataset.scala:3316) at org.apache.spark.sql.Dataset.$anonfun$withAction$2(Dataset.scala:4323) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546) at org.apache.spark.sql.Dataset.$anonfun$withAction$1(Dataset.scala:4321) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:201) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:108) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:66) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:4321) at org.apache.spark.sql.Dataset.head(Dataset.scala:3316) at org.apache.spark.sql.Dataset.take(Dataset.scala:3539) at org.apache.spark.sql.Dataset.takeAsList(Dataset.scala:3562) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.take(DataFrameCommonJob.java:145) at com.adb.hdh.spark.job.impl.DataFrameCommonJob.checkSqlDb(DataFrameCommonJob.java:276) at com.adb.hdh.spark.jobs.SparkJobRunnerService.runJob(SparkJobRunnerService.java:183) at com.adb.hdh.spark.jobs.AbstractJobThread.run(AbstractJobThread.java:47) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) at java.base/java.util.concurrent.FutureTask.run(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) |
|