digraph G {
0 [labelType="html" label="<br><b>AdaptiveSparkPlan</b><br><br>"];
subgraph cluster1 {
isCluster="true";
label="WholeStageCodegen (2)\n \nduration: 1 ms";
2 [labelType="html" label="<b>HashAggregate</b><br><br>time in aggregation build: 0 ms<br>number of output rows: 1"];
}
3 [labelType="html" label="<b>Exchange</b><br><br>shuffle records written: 5<br>local merged chunks fetched: 0<br>shuffle write time total (min, med, max (stageId: taskId))<br>3 ms (0 ms, 0 ms, 0 ms (stage 903.0: task 1233))<br>remote merged bytes read: 0.0 B<br>local merged blocks fetched: 0<br>corrupt merged block chunks: 0<br>remote merged reqs duration: 0 ms<br>remote merged blocks fetched: 0<br>records read: 5<br>local bytes read: 295.0 B<br>fetch wait time: 0 ms<br>remote bytes read: 0.0 B<br>merged fetch fallback count: 0<br>local blocks read: 5<br>remote merged chunks fetched: 0<br>remote blocks read: 0<br>data size total (min, med, max (stageId: taskId))<br>80.0 B (16.0 B, 16.0 B, 16.0 B (stage 903.0: task 1232))<br>local merged bytes read: 0.0 B<br>number of partitions: 1<br>remote reqs duration: 0 ms<br>remote bytes read to disk: 0.0 B<br>shuffle bytes written total (min, med, max (stageId: taskId))<br>295.0 B (59.0 B, 59.0 B, 59.0 B (stage 903.0: task 1232))"];
subgraph cluster4 {
isCluster="true";
label="WholeStageCodegen (1)\n \nduration: total (min, med, max (stageId: taskId))\n15 ms (3 ms, 3 ms, 3 ms (stage 903.0: task 1232))";
5 [labelType="html" label="<b>HashAggregate</b><br><br>time in aggregation build total (min, med, max (stageId: taskId))<br>14 ms (2 ms, 3 ms, 3 ms (stage 903.0: task 1232))<br>number of output rows: 5"];
}
6 [labelType="html" label="<b>Scan binaryFile </b><br><br>number of output rows: 5<br>number of files read: 5<br>metadata time: 0 ms<br>size of files read: 34.6 MiB"];
2->0;
3->2;
5->3;
6->5;
}
7
AdaptiveSparkPlan isFinalPlan=true
HashAggregate(keys=[], functions=[count(1)])
WholeStageCodegen (2)
Exchange SinglePartition, ENSURE_REQUIREMENTS, [plan_id=14922]
HashAggregate(keys=[], functions=[partial_count(1)])
WholeStageCodegen (1)
FileScan binaryFile [] Batched: false, DataFilters: [], Format: org.apache.spark.sql.execution.datasources.binaryfile.BinaryFileFormat@4ecbe813, Location: InMemoryFileIndex(5 paths)[file:/data/input/depot/binary/execution/454FFE78_6FB4_4FFA_AB5C_590D50..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<>
== Physical Plan ==
AdaptiveSparkPlan (9)
+- == Final Plan ==
* HashAggregate (5)
+- ShuffleQueryStage (4), Statistics(sizeInBytes=80.0 B, rowCount=5)
+- Exchange (3)
+- * HashAggregate (2)
+- Scan binaryFile (1)
+- == Initial Plan ==
HashAggregate (8)
+- Exchange (7)
+- HashAggregate (6)
+- Scan binaryFile (1)
(1) Scan binaryFile
Output: []
Batched: false
Location: InMemoryFileIndex [file:/data/input/depot/binary/execution/454FFE78_6FB4_4FFA_AB5C_590D50B78C20/current/BENARD_Clement_these_2021.pdf, ... 4 entries]
ReadSchema: struct<>
(2) HashAggregate [codegen id : 1]
Input: []
Keys: []
Functions [1]: [partial_count(1)]
Aggregate Attributes [1]: [count#147316L]
Results [1]: [count#147317L]
(3) Exchange
Input [1]: [count#147317L]
Arguments: SinglePartition, ENSURE_REQUIREMENTS, [plan_id=14922]
(4) ShuffleQueryStage
Output [1]: [count#147317L]
Arguments: 0
(5) HashAggregate [codegen id : 2]
Input [1]: [count#147317L]
Keys: []
Functions [1]: [count(1)]
Aggregate Attributes [1]: [count(1)#147313L]
Results [1]: [count(1)#147313L AS count#147314L]
(6) HashAggregate
Input: []
Keys: []
Functions [1]: [partial_count(1)]
Aggregate Attributes [1]: [count#147316L]
Results [1]: [count#147317L]
(7) Exchange
Input [1]: [count#147317L]
Arguments: SinglePartition, ENSURE_REQUIREMENTS, [plan_id=14914]
(8) HashAggregate
Input [1]: [count#147317L]
Keys: []
Functions [1]: [count(1)]
Aggregate Attributes [1]: [count(1)#147313L]
Results [1]: [count(1)#147313L AS count#147314L]
(9) AdaptiveSparkPlan
Output [1]: [count#147314L]
Arguments: isFinalPlan=true