digraph G {
0 [labelType="html" label="<br><b>AdaptiveSparkPlan</b><br><br>"];
subgraph cluster1 {
isCluster="true";
label="WholeStageCodegen (5)\n \nduration: 0 ms";
2 [labelType="html" label="<b>HashAggregate</b><br><br>time in aggregation build: 0 ms<br>number of output rows: 1"];
}
3 [labelType="html" label="<b>Exchange</b><br><br>shuffle records written: 4<br>local merged chunks fetched: 0<br>shuffle write time total (min, med, max (stageId: taskId))<br>2 ms (0 ms, 0 ms, 0 ms (stage 507.0: task 408))<br>remote merged bytes read: 0.0 B<br>local merged blocks fetched: 0<br>corrupt merged block chunks: 0<br>remote merged reqs duration: 0 ms<br>remote merged blocks fetched: 0<br>records read: 4<br>local bytes read: 233.0 B<br>fetch wait time: 0 ms<br>remote bytes read: 0.0 B<br>merged fetch fallback count: 0<br>local blocks read: 4<br>remote merged chunks fetched: 0<br>remote blocks read: 0<br>data size total (min, med, max (stageId: taskId))<br>64.0 B (16.0 B, 16.0 B, 16.0 B (stage 507.0: task 408))<br>local merged bytes read: 0.0 B<br>number of partitions: 1<br>remote reqs duration: 0 ms<br>remote bytes read to disk: 0.0 B<br>shuffle bytes written total (min, med, max (stageId: taskId))<br>233.0 B (56.0 B, 59.0 B, 59.0 B (stage 507.0: task 410))"];
subgraph cluster4 {
isCluster="true";
label="WholeStageCodegen (4)\n \nduration: total (min, med, max (stageId: taskId))\n18 ms (3 ms, 5 ms, 5 ms (stage 507.0: task 410))";
5 [labelType="html" label="<b>HashAggregate</b><br><br>time in aggregation build total (min, med, max (stageId: taskId))<br>17 ms (3 ms, 5 ms, 5 ms (stage 507.0: task 409))<br>number of output rows: 4"];
}
6 [labelType="html" label="<br><b>Union</b><br><br>"];
7 [labelType="html" label="<b>Scan csv </b><br><br>number of output rows: 0<br>number of files read: 1<br>metadata time: 0 ms<br>size of files read: 2.0 B"];
subgraph cluster8 {
isCluster="true";
label="WholeStageCodegen (1)\n \nduration: 15 ms";
9 [labelType="html" label="<b>ColumnarToRow</b><br><br>number of output rows: 25<br>number of input batches: 1"];
}
10 [labelType="html" label="<b>Scan parquet </b><br><br>number of files read: 1<br>scan time: 5 ms<br>metadata time: 0 ms<br>size of files read: 52.4 KiB<br>number of output rows: 25"];
subgraph cluster11 {
isCluster="true";
label="WholeStageCodegen (2)\n \nduration: 14 ms";
12 [labelType="html" label="<b>ColumnarToRow</b><br><br>number of output rows: 6<br>number of input batches: 1"];
}
13 [labelType="html" label="<b>Scan parquet </b><br><br>number of files read: 1<br>scan time: 4 ms<br>metadata time: 0 ms<br>size of files read: 35.3 KiB<br>number of output rows: 6"];
subgraph cluster14 {
isCluster="true";
label="WholeStageCodegen (3)\n \nduration: 6 ms";
15 [labelType="html" label="<b>ColumnarToRow</b><br><br>number of output rows: 32<br>number of input batches: 1"];
}
16 [labelType="html" label="<b>Scan parquet </b><br><br>number of files read: 1<br>scan time: 4 ms<br>metadata time: 0 ms<br>size of files read: 59.9 KiB<br>number of output rows: 32"];
2->0;
3->2;
5->3;
6->5;
7->6;
9->6;
10->9;
12->6;
13->12;
15->6;
16->15;
}
17
AdaptiveSparkPlan isFinalPlan=true
HashAggregate(keys=[], functions=[count(1)])
WholeStageCodegen (5)
Exchange SinglePartition, ENSURE_REQUIREMENTS, [plan_id=13235]
HashAggregate(keys=[], functions=[partial_count(1)])
WholeStageCodegen (4)
Union
FileScan csv [] Batched: false, DataFilters: [], Format: CSV, Location: InMemoryFileIndex(1 paths)[file:/data/input/depot/csv/execution/empty.csv], PartitionFilters: [], PushedFilters: [], ReadSchema: struct<>
ColumnarToRow
WholeStageCodegen (1)
FileScan parquet [] Batched: true, DataFilters: [], Format: Parquet, Location: InMemoryFileIndex(1 paths)[file:/data/input/depot/parquet/execution/VLIB_C_API_QUALITY_REPORT/cur..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<>
ColumnarToRow
WholeStageCodegen (2)
FileScan parquet [] Batched: true, DataFilters: [], Format: Parquet, Location: InMemoryFileIndex(1 paths)[file:/data/input/depot/parquet/execution/TEST_VENDRED_C_API_QUALITY_RE..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<>
ColumnarToRow
WholeStageCodegen (3)
FileScan parquet [] Batched: true, DataFilters: [], Format: Parquet, Location: InMemoryFileIndex(1 paths)[file:/data/input/depot/parquet/execution/MED_GIYGA_C_API_QUALITY_REPOR..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<>