You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/declarative-pipelines/PipelineExecution.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,13 +18,13 @@
18
18
runPipeline():Unit
19
19
```
20
20
21
-
`runPipeline`[starts the pipeline](#startPipeline) and requests the [PipelineExecution](PipelineUpdateContext.md#pipelineExecution) (of this [PipelineUpdateContext](#context)) to [wait for the execution to complete](#awaitCompletion).
21
+
`runPipeline`[starts this pipeline](#startPipeline) and requests the [PipelineExecution](PipelineUpdateContext.md#pipelineExecution) (of this [PipelineUpdateContext](#context)) to [wait for the execution to complete](#awaitCompletion).
22
22
23
23
---
24
24
25
25
`runPipeline` is used when:
26
26
27
-
*`PipelinesHandler` is requested to [startRun](PipelinesHandler.md#startRun) (for [Spark Connect]({{ book.spark_connect }}))
27
+
*`PipelinesHandler` is requested to [start a pipeline run](PipelinesHandler.md#startRun)
Copy file name to clipboardExpand all lines: docs/declarative-pipelines/PipelinesHandler.md
+35-22Lines changed: 35 additions & 22 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,6 +2,8 @@
2
2
3
3
`PipelinesHandler` is used to [handle pipeline commands](#handlePipelinesCommand) in [Spark Connect]({{ book.spark_connect }}) ([SparkConnectPlanner]({{ book.spark_connect }}/server/SparkConnectPlanner), precisely).
4
4
5
+
`PipelinesHandler` acts as a bridge between Python and SQL "frontends" and Spark Connect Server (where pipeline execution happens).
`handlePipelinesCommand`[defines a flow](#defineFlow).
57
63
58
-
### Start Pipeline { #startRun }
64
+
### START_RUN { #START_RUN }
65
+
66
+
`handlePipelinesCommand` prints out the following INFO message to the logs:
67
+
68
+
```text
69
+
Start pipeline cmd received: [cmd]
70
+
```
71
+
72
+
`handlePipelinesCommand`[starts a pipeline run](#startRun).
73
+
74
+
## Start Pipeline Run { #startRun }
59
75
60
76
```scala
61
77
startRun(
@@ -64,21 +80,18 @@ startRun(
64
80
sessionHolder: SessionHolder):Unit
65
81
```
66
82
67
-
`startRun` prints out the following INFO message to the logs:
68
-
69
-
```text
70
-
Start pipeline cmd received: [cmd]
71
-
```
83
+
??? note "`START_RUN` Pipeline Command"
84
+
`startRun` is used when `PipelinesHandler` is requested to handle [proto.PipelineCommand.CommandTypeCase.START_RUN](#START_RUN) command.
72
85
73
86
`startRun` finds the [GraphRegistrationContext](GraphRegistrationContext.md) by `dataflowGraphId` in the [DataflowGraphRegistry](DataflowGraphRegistry.md) (in the given `SessionHolder`).
74
87
75
88
`startRun` creates a `PipelineEventSender` to send pipeline events back to the Spark Connect client (_Python pipeline runtime_).
76
89
77
90
`startRun` creates a [PipelineUpdateContextImpl](PipelineUpdateContextImpl.md) (with the `PipelineEventSender`).
78
91
79
-
In the end, `startRun` requests the `PipelineUpdateContextImpl` for the [PipelineExecution](PipelineExecution.md) to [runPipeline](PipelineExecution.md#runPipeline) or [dryRunPipeline](PipelineExecution.md#dryRunPipeline) for `dry-run` or `run` command, respectively.
92
+
In the end, `startRun` requests the `PipelineUpdateContextImpl` for the [PipelineExecution](PipelineUpdateContext.md#pipelineExecution) to [run a pipeline](PipelineExecution.md#runPipeline) or [dry-run a pipeline](PipelineExecution.md#dryRunPipeline) for `dry-run` or `run` command, respectively.
80
93
81
-
###Create Dataflow Graph { #createDataflowGraph }
94
+
## Create Dataflow Graph { #createDataflowGraph }
82
95
83
96
```scala
84
97
createDataflowGraph(
@@ -90,7 +103,7 @@ createDataflowGraph(
90
103
91
104
`createDataflowGraph` returns the ID of the created dataflow graph.
0 commit comments