Skip to content

Commit 5a2ba02

Browse files
Fix a few typos (#233)
1 parent e0a73ce commit 5a2ba02

File tree

3 files changed

+5
-5
lines changed

3 files changed

+5
-5
lines changed

01x_lazy.ipynb

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -214,7 +214,7 @@
214214
"cell_type": "markdown",
215215
"metadata": {},
216216
"source": [
217-
"Your task is to recreate this graph again using the delayed function on the original Python code. The three functions you want to delay are `pd.read_csv`, `len` and `sum`.. "
217+
"Your task is to recreate this graph again using the delayed function on the original Python code. The three functions you want to delay are `pd.read_csv`, `len` and `sum`. "
218218
]
219219
},
220220
{
@@ -558,7 +558,7 @@
558558
"source": [
559559
"Any Dask object, such as `total`, above, has an attribute which describes the calculations necessary to produce that result. Indeed, this is exactly the graph that we have been talking about, which can be visualized. We see that it is a simple dictionary, in which the keys are unique task identifiers, and the values are the functions and inputs for calculation.\n",
560560
"\n",
561-
"`delayed` is a handy mechanism for creating the Dask graph, but the adventurous may wish to play with the full fexibility afforded by building the graph dictionaries directly. Detailed information can be found [here](http://dask.pydata.org/en/latest/graphs.html)."
561+
"`delayed` is a handy mechanism for creating the Dask graph, but the adventurous may wish to play with the full flexibility afforded by building the graph dictionaries directly. Detailed information can be found [here](http://dask.pydata.org/en/latest/graphs.html)."
562562
]
563563
},
564564
{

06_distributed_advanced.ipynb

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -630,10 +630,10 @@
630630
"metadata": {},
631631
"source": [
632632
"The trouble with this approach is that Dask is meant for the execution of large datasets/computations - you probably can't simply run the whole thing \n",
633-
"in one local thread, else you wouldn't have used Dask in the first place. So the code above should only be used on a small part of the data that also exihibits the error. \n",
633+
"in one local thread, else you wouldn't have used Dask in the first place. So the code above should only be used on a small part of the data that also exhibits the error. \n",
634634
"Furthermore, the method will not work when you are dealing with futures (such as `f`, above, or after persisting) instead of delayed-based computations.\n",
635635
"\n",
636-
"As an alternative, you can ask the scheduler to analyze your calculation and find the specific sub-task responsible for the error, and pull only it and its dependnecies locally for execution."
636+
"As an alternative, you can ask the scheduler to analyze your calculation and find the specific sub-task responsible for the error, and pull only it and its dependencies locally for execution."
637637
]
638638
},
639639
{

08_machine_learning.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -243,7 +243,7 @@
243243
"\n",
244244
"Dask can talk to scikit-learn (via joblib) so that your *cluster* is used to train a model. \n",
245245
"\n",
246-
"If you run this on a laptop, it will take quite some time, but the CPU usage will be satisfyingly near 100% for the duration. To run faster, you would need a disrtibuted cluster. That would mean putting something in the call to `Client` something like\n",
246+
"If you run this on a laptop, it will take quite some time, but the CPU usage will be satisfyingly near 100% for the duration. To run faster, you would need a distributed cluster. That would mean putting something in the call to `Client` something like\n",
247247
"\n",
248248
"```\n",
249249
"c = Client('tcp://my.scheduler.address:8786')\n",

0 commit comments

Comments
 (0)