You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
3. There is a small data cube included in the repo that is used in the tutorials. We don't want git to track this so we tell it to ignore this file path.
16
+
3. There is a small data cube included in the repo that is used in the tutorials. We don't want git to track this so we tell it to ignore this file path.
Note that the first `pixi run` will download specific versions of all required Python libraries to a hidden directory `./.pixi`. Subsequent runs activate that environment and execute code within it. You can also run `pixi shell` to "activate" the environment (set paths to executables and auxiliary files) and `exit` to deactivate it.
24
24
@@ -27,16 +27,17 @@ Note that the first `pixi run` will download specific versions of all required P
3. Create and activate a conda environment from the `environment.yml` file located in the repo:
34
-
```conda env create -f environment.yml```
34
+
```conda env create -f environment.yml```
35
+
```conda activate book```
35
36
36
-
4. There is a small data cube included in the repo that is used in the tutorials. We don't want git to track this so we tell it to ignore this file path.
37
+
4. There is a small data cube included in the repo that is used in the tutorials. We don't want git to track this so we tell it to ignore this file path.
5. Start Jupyterlab and navigate to the directories containing the Jupyter notebooks (`itslive/nbs` and `s1/nbs`):
40
+
5. Start Jupyterlab and navigate to the directories containing the Jupyter notebooks (`itslive/nbs` and `s1/nbs`):
40
41
```jupyterlab```
41
42
42
43
Both tutorials use functions that are stored in scripts associated with each dataset. You can find these scripts here: [`itslive_tools.py`](../itslive/nbs/itslive_tools.py) and [`s1_tools.py`](../sentinel1/nbs/s1_tools.py).
Copy file name to clipboardExpand all lines: book/itslive/nbs/2_larger_than_memory_data.ipynb
+1-3Lines changed: 1 addition & 3 deletions
Original file line number
Diff line number
Diff line change
@@ -547,9 +547,7 @@
547
547
"source": [
548
548
"{{conclusion}}\n",
549
549
"\n",
550
-
"In this notebook, we identified a strategy for reading, chunking, and organizing this dataset that works within the memory constraints of my laptop and the size of the data. \n",
551
-
"\n",
552
-
"In the next notebook, we use vector data to narrow our focus in on a spatial area of interest and start examining ice velocity data."
550
+
"In this notebook, we identified a strategy for reading, chunking, and organizing this dataset that works within the memory constraints of my laptop and the size of the data. In the next notebook, we use vector data to narrow our focus in on a spatial area of interest and start examining ice velocity data."
"This notebook demonstrated reading large data into memory by creating a virtual dataset that references that full dataset without directly reading it. \n",
509
-
"\n",
510
-
"However, we also saw that reading the data in this way produces an object that lacks important metadata. The next notebook will go through the steps of locating and adding relevant metadata to the backscatter data cubes read in this notebook.\n"
508
+
"This notebook demonstrated reading large data into memory by creating a virtual dataset that references that full dataset without directly reading it. However, we also saw that reading the data in this way produces an object that lacks important metadata. The next notebook will go through the steps of locating and adding relevant metadata to the backscatter data cubes read in this notebook.\n"
0 commit comments