digraph G {
0 [labelType="html" label="<b>Execute InsertIntoHadoopFsRelationCommand</b><br><br>task commit time: 13 ms<br>number of written files: 1<br>job commit time: 19 ms<br>number of output rows: 82<br>number of dynamic part: 0<br>written output: 7.5 KiB"];
1 [labelType="html" label="<br><b>WriteFiles</b><br><br>"];
subgraph cluster2 {
isCluster="true";
label="WholeStageCodegen (1)\n \nduration: 139 ms";
3 [labelType="html" label="<br><b>Project</b><br><br>"];
4 [labelType="html" label="<b>Scan JDBCRelation((SELECT \"VISIT_ID\", \"PATIENT_ID\", \"ID_SEJ\", \"ID_PAT\", \"IN_DATE\", \"OUT_DATE\", \"IN_MODE\", \"OUT_MODE\", \"IN_UNIT\", \"OUT_UNIT\", \"VISIT_TYPE\", \"CHAMP_PMSI\", \"UNIT_PATH\", \"EMERGENCY\", \"CANCELED\" FROM \"DTM_261\".\"VISIT\") x) [numPartitions=1] </b><br><br>number of output rows: 82"];
}
1->0;
3->1;
4->3;
}
5
Execute InsertIntoHadoopFsRelationCommand file:/data/output/cache/parquet/uet/VISIT_2941, false, Parquet, [path=file:///data/output/cache/parquet/uet/VISIT_2941], Overwrite, [CANCELED_2956, CHAMP_PMSI_2953, EMERGENCY_2955, ID_PAT_2945, ID_SEJ_2944, IN_DATE_2946, IN_MODE_2948, IN_UNIT_2950, OUT_DATE_2947, OUT_MODE_2949, OUT_UNIT_2951, PATIENT_ID_2943, UNIT_PATH_2954, VISIT_ID_2942, VISIT_TYPE_2952]
WriteFiles
Project [cast(CANCELED#3486 as int) AS CANCELED_2956#3578, CHAMP_PMSI#3483 AS CHAMP_PMSI_2953#3579, staticinvoke(class org.apache.spark.sql.catalyst.util.CharVarcharCodegenUtils, StringType, readSidePadding, EMERGENCY#3485, 1, true, false, true) AS EMERGENCY_2955#3580, cast(ID_PAT#3475 as int) AS ID_PAT_2945#3581, cast(ID_SEJ#3474 as int) AS ID_SEJ_2944#3582, IN_DATE#3476 AS IN_DATE_2946#3583, IN_MODE#3478 AS IN_MODE_2948#3584, IN_UNIT#3480 AS IN_UNIT_2950#3585, OUT_DATE#3477 AS OUT_DATE_2947#3586, OUT_MODE#3479 AS OUT_MODE_2949#3587, OUT_UNIT#3481 AS OUT_UNIT_2951#3588, null AS PATIENT_ID_2943#3589, UNIT_PATH#3484 AS UNIT_PATH_2954#3590, null AS VISIT_ID_2942#3591, VISIT_TYPE#3482 AS VISIT_TYPE_2952#3592]
Scan JDBCRelation((SELECT "VISIT_ID", "PATIENT_ID", "ID_SEJ", "ID_PAT", "IN_DATE", "OUT_DATE", "IN_MODE", "OUT_MODE", "IN_UNIT", "OUT_UNIT", "VISIT_TYPE", "CHAMP_PMSI", "UNIT_PATH", "EMERGENCY", "CANCELED" FROM "DTM_261"."VISIT") x) [numPartitions=1] [CANCELED#3486,CHAMP_PMSI#3483,EMERGENCY#3485,ID_PAT#3475,ID_SEJ#3474,IN_DATE#3476,IN_MODE#3478,IN_UNIT#3480,OUT_DATE#3477,OUT_MODE#3479,OUT_UNIT#3481,UNIT_PATH#3484,VISIT_TYPE#3482] PushedFilters: [], ReadSchema: struct<CANCELED:decimal(1,0),CHAMP_PMSI:string,EMERGENCY:string,ID_PAT:decimal(38,0),ID_SEJ:decim...
WholeStageCodegen (1)
== Physical Plan ==
Execute InsertIntoHadoopFsRelationCommand (4)
+- WriteFiles (3)
+- * Project (2)
+- * Scan JDBCRelation((SELECT "VISIT_ID", "PATIENT_ID", "ID_SEJ", "ID_PAT", "IN_DATE", "OUT_DATE", "IN_MODE", "OUT_MODE", "IN_UNIT", "OUT_UNIT", "VISIT_TYPE", "CHAMP_PMSI", "UNIT_PATH", "EMERGENCY", "CANCELED" FROM "DTM_261"."VISIT") x) [numPartitions=1] (1)
(1) Scan JDBCRelation((SELECT "VISIT_ID", "PATIENT_ID", "ID_SEJ", "ID_PAT", "IN_DATE", "OUT_DATE", "IN_MODE", "OUT_MODE", "IN_UNIT", "OUT_UNIT", "VISIT_TYPE", "CHAMP_PMSI", "UNIT_PATH", "EMERGENCY", "CANCELED" FROM "DTM_261"."VISIT") x) [numPartitions=1] [codegen id : 1]
Output [13]: [CANCELED#3486, CHAMP_PMSI#3483, EMERGENCY#3485, ID_PAT#3475, ID_SEJ#3474, IN_DATE#3476, IN_MODE#3478, IN_UNIT#3480, OUT_DATE#3477, OUT_MODE#3479, OUT_UNIT#3481, UNIT_PATH#3484, VISIT_TYPE#3482]
ReadSchema: struct<CANCELED:decimal(1,0),CHAMP_PMSI:string,EMERGENCY:string,ID_PAT:decimal(38,0),ID_SEJ:decimal(38,0),IN_DATE:timestamp,IN_MODE:string,IN_UNIT:string,OUT_DATE:timestamp,OUT_MODE:string,OUT_UNIT:string,UNIT_PATH:string,VISIT_TYPE:string>
(2) Project [codegen id : 1]
Output [15]: [cast(CANCELED#3486 as int) AS CANCELED_2956#3578, CHAMP_PMSI#3483 AS CHAMP_PMSI_2953#3579, staticinvoke(class org.apache.spark.sql.catalyst.util.CharVarcharCodegenUtils, StringType, readSidePadding, EMERGENCY#3485, 1, true, false, true) AS EMERGENCY_2955#3580, cast(ID_PAT#3475 as int) AS ID_PAT_2945#3581, cast(ID_SEJ#3474 as int) AS ID_SEJ_2944#3582, IN_DATE#3476 AS IN_DATE_2946#3583, IN_MODE#3478 AS IN_MODE_2948#3584, IN_UNIT#3480 AS IN_UNIT_2950#3585, OUT_DATE#3477 AS OUT_DATE_2947#3586, OUT_MODE#3479 AS OUT_MODE_2949#3587, OUT_UNIT#3481 AS OUT_UNIT_2951#3588, null AS PATIENT_ID_2943#3589, UNIT_PATH#3484 AS UNIT_PATH_2954#3590, null AS VISIT_ID_2942#3591, VISIT_TYPE#3482 AS VISIT_TYPE_2952#3592]
Input [13]: [CANCELED#3486, CHAMP_PMSI#3483, EMERGENCY#3485, ID_PAT#3475, ID_SEJ#3474, IN_DATE#3476, IN_MODE#3478, IN_UNIT#3480, OUT_DATE#3477, OUT_MODE#3479, OUT_UNIT#3481, UNIT_PATH#3484, VISIT_TYPE#3482]
(3) WriteFiles
Input [15]: [CANCELED_2956#3578, CHAMP_PMSI_2953#3579, EMERGENCY_2955#3580, ID_PAT_2945#3581, ID_SEJ_2944#3582, IN_DATE_2946#3583, IN_MODE_2948#3584, IN_UNIT_2950#3585, OUT_DATE_2947#3586, OUT_MODE_2949#3587, OUT_UNIT_2951#3588, PATIENT_ID_2943#3589, UNIT_PATH_2954#3590, VISIT_ID_2942#3591, VISIT_TYPE_2952#3592]
(4) Execute InsertIntoHadoopFsRelationCommand
Input: []
Arguments: file:/data/output/cache/parquet/uet/VISIT_2941, false, Parquet, [path=file:///data/output/cache/parquet/uet/VISIT_2941], Overwrite, [CANCELED_2956, CHAMP_PMSI_2953, EMERGENCY_2955, ID_PAT_2945, ID_SEJ_2944, IN_DATE_2946, IN_MODE_2948, IN_UNIT_2950, OUT_DATE_2947, OUT_MODE_2949, OUT_UNIT_2951, PATIENT_ID_2943, UNIT_PATH_2954, VISIT_ID_2942, VISIT_TYPE_2952]