You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/synapse-analytics/get-started.md
+13-4Lines changed: 13 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -103,7 +103,9 @@ Once your Synapse workspace is created, you have two ways to open Synapse Studio
103
103
|---|---|---|
104
104
|**Apache Spark pool name**|`Spark1`
105
105
|**Node size**|`Small`|
106
-
|**Number of nodes**| Set the minimum to 3 and the maximum to 3
106
+
|**Number of nodes**| Set the minimum to 3 and the maximum to 3|
107
+
|||
108
+
107
109
* Select **Review+create** and then select **Create**.
108
110
* Your Apache Spark pool will be ready in a few seconds.
109
111
@@ -116,6 +118,7 @@ Once your Synapse workspace is created, you have two ways to open Synapse Studio
116
118
> Spark databases are independently created from Spark pools. A workspace always has a Spark DB called **default** and you can create additional Spark databases.
117
119
118
120
## SQL on-demand pools
121
+
119
122
SQL on-demand is a special kind of SQL pool that is always available with a Synapse workspace. It allows you to work with SQL without having to create or think about managing a Synapse SQL pool.
120
123
121
124
> [!NOTE]
@@ -208,7 +211,8 @@ We have data available in a SQL pool database. Now we load it into a Spark datab
208
211
## Customize data visualization data with Spark and notebooks
209
212
210
213
With spark notebooks you can control exactly how render charts. The following
211
-
code shows a simple example using the popular libraries matplotlib and sea-born.
214
+
code shows a simple example using the popular libraries matplotlib and sea-born. It will
215
+
render the same chart you saw when running the SQL queries earlier.
212
216
213
217
```py
214
218
%%pyspark
@@ -223,7 +227,12 @@ seaborn.lineplot(x="PassengerCount", y="AvgTripDistance" , data = df)
223
227
matplotlib.pyplot.show()
224
228
```
225
229
226
-
## Load data from a Spark table into a SQL Pool table
230
+
## Load data from a Spark table into a SQL pool table
231
+
232
+
Earlier we copied data from a SQL pool database into a Spark DB. Using
233
+
Spark, we aggregated the data into the nyctaxi.passengercountstats.
234
+
Now run the cell below in a notebook and it will copy the aggregated table back into
* NOTE: THe first time you run this it will take about 10 seconds for SQL on-demand to gather SQL resources needed to run your queries. Every subsequent query will not require this time.
256
+
* NOTE: THe first time you run this it will take about 10 seconds for SQL on-demand to gather SQL resources needed to run your queries. Subsequent queries will not require this time.
0 commit comments