|
116 | 116 | " \n",
|
117 | 117 | "Here comes the tricky part!\n",
|
118 | 118 | " \n",
|
119 |
| - "The input to the pipeline will be our dataframe `X`, which one row per identifier.\n", |
| 119 | + "The input to the pipeline will be our dataframe `X`, with one row per identifier.\n", |
120 | 120 | "It is currently empty.\n",
|
121 |
| - "But which time series data should the `RelevantFeatureAugmenter` to actually extract the features from?\n", |
| 121 | + "But which time series data should the `RelevantFeatureAugmenter` use to actually extract the features from?\n", |
122 | 122 | "\n",
|
123 | 123 | "We need to pass the time series data (stored in `df_ts`) to the transformer.\n",
|
124 | 124 | " \n",
|
|
179 | 179 | "cell_type": "markdown",
|
180 | 180 | "metadata": {},
|
181 | 181 | "source": [
|
182 |
| - "During interference, the augmentor does only extract the relevant features it has found out in the training phase and the classifier predicts the target using these features." |
| 182 | + "During inference, the augmenter only extracts those features that it has found as being relevant in the training phase. The classifier predicts the target using these features." |
183 | 183 | ]
|
184 | 184 | },
|
185 | 185 | {
|
|
211 | 211 | "cell_type": "markdown",
|
212 | 212 | "metadata": {},
|
213 | 213 | "source": [
|
214 |
| - "You can also find out, which columns the augmenter has selected" |
| 214 | + "You can also find out which columns the augmenter has selected" |
215 | 215 | ]
|
216 | 216 | },
|
217 | 217 | {
|
|
248 | 248 | "metadata": {},
|
249 | 249 | "source": [
|
250 | 250 | "In the example above we passed in a single `df_ts` into the `RelevantFeatureAugmenter`, which was used both for training and predicting.\n",
|
251 |
| - "During training, only the data with the `id`s from `X_train` where extracted and during prediction the rest.\n", |
| 251 | + "During training, only the data with the `id`s from `X_train` were extracted. The rest of the data are extracted during prediction.\n", |
252 | 252 | "\n",
|
253 | 253 | "However, it is perfectly fine to call `set_params` twice: once before training and once before prediction. \n",
|
254 | 254 | "This can be handy if you for example dump the trained pipeline to disk and re-use it only later for prediction.\n",
|
255 |
| - "You only need to make sure that the `id`s of the enteties you use during training/prediction are actually present in the passed time series data." |
| 255 | + "You only need to make sure that the `id`s of the entities you use during training/prediction are actually present in the passed time series data." |
256 | 256 | ]
|
257 | 257 | },
|
258 | 258 | {
|
|
0 commit comments