Skip to content

Commit c6fe18c

Browse files
pcoetcopybara-github
authored andcommitted
Fix up CSV data tutorial: fix typos; link to pandas.DataFrame docs on first usage, rather than second; add transition sentence to mixed data types section.
PiperOrigin-RevId: 572711158
1 parent 2bcfa24 commit c6fe18c

File tree

1 file changed

+7
-5
lines changed

1 file changed

+7
-5
lines changed

site/en/tutorials/load_data/csv.ipynb

Lines changed: 7 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -120,7 +120,7 @@
120120
"id": "ny5TEgcmHjVx"
121121
},
122122
"source": [
123-
"For any small CSV dataset the simplest way to train a TensorFlow model on it is to load it into memory as a pandas Dataframe or a NumPy array.\n"
123+
"For any small CSV dataset the simplest way to train a TensorFlow model on it is to load it into memory as a [pandas `DataFrame`](https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.html) or a NumPy array.\n"
124124
]
125125
},
126126
{
@@ -132,9 +132,9 @@
132132
"A relatively simple example is the [abalone dataset](https://archive.ics.uci.edu/ml/datasets/abalone).\n",
133133
"\n",
134134
"* The dataset is small.\n",
135-
"* All the input features are all limited-range floating point values.\n",
135+
"* All the input features are limited-range floating point values.\n",
136136
"\n",
137-
"Here is how to download the data into a [pandas `DataFrame`](https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.html):"
137+
"Here is how to download the data into a `DataFrame`:"
138138
]
139139
},
140140
{
@@ -355,6 +355,8 @@
355355
"source": [
356356
"## Mixed data types\n",
357357
"\n",
358+
"In the previous sections, you worked with a dataset where all the features were limited-range floating point values. But not all datasets are limited to a single data type.\n",
359+
"\n",
358360
"The \"Titanic\" dataset contains information about the passengers on the Titanic. The nominal task on this dataset is to predict who survived.\n",
359361
"\n",
360362
"![The Titanic](images/csv/Titanic.jpg)\n",
@@ -903,7 +905,7 @@
903905
"source": [
904906
"### From a single file\n",
905907
"\n",
906-
"So far this tutorial has worked with in-memory data. `tf.data` is a highly scalable toolkit for building data pipelines, and provides a few functions for dealing loading CSV files. "
908+
"So far this tutorial has worked with in-memory data. `tf.data` is a highly scalable toolkit for building data pipelines, and provides a few functions for loading CSV files. "
907909
]
908910
},
909911
{
@@ -1373,7 +1375,7 @@
13731375
"id": "3jiGZeUijJNd"
13741376
},
13751377
"source": [
1376-
"So far this tutorial has focused on the highest-level utilities for reading csv data. There are other two APIs that may be helpful for advanced users if your use-case doesn't fit the basic patterns.\n",
1378+
"So far this tutorial has focused on the highest-level utilities for reading csv data. There are two other APIs that may be helpful for advanced users if your use-case doesn't fit the basic patterns.\n",
13771379
"\n",
13781380
"* `tf.io.decode_csv`: a function for parsing lines of text into a list of CSV column tensors.\n",
13791381
"* `tf.data.experimental.CsvDataset`: a lower-level CSV dataset constructor.\n",

0 commit comments

Comments
 (0)