@@ -261,16 +261,17 @@ For these kinds of tasks: usually use one-hot-encoding to predict the
261
261
by column for speed and more local correlation.
262
262
263
263
Number of classes when reading...
264
- - ... column by column: @@beamer:$\approx 17\,600$ \uncover<2->{\alert{digit} number}@@
265
- - ... tile by tile: 662
264
+ - ... column by column: @@beamer:$\approx 17\,600$
265
+ \uncover<2->{\alert{digit} number}@@
266
+ - ... tile by tile: 662@@beamer: \uncover<2->{\alert{digit} number}@@
266
267
\pause No way I could train a model on that many classes even if
267
268
memory problems were solved.
268
269
*** Dimensionalities
269
270
Different complexities reflected on the highest level via
270
271
dimensionality:
271
272
- 1D: Level is seen as single row of one type of tile.
272
273
- 2D: Level is seen as matrix of one type of tile.
273
- - 3D: Level is seen as cube with chosen tile layers.
274
+ - 3D: Level is seen as cuboid with chosen tile layers.
274
275
The default type of tile for 1D and 2D is the ground tile of the
275
276
level.
276
277
*** Simplifications
@@ -360,15 +361,16 @@ manually chosen stride, padding and dilation so output size matches
360
361
first screen size.
361
362
362
363
Discriminators: Inputs are first screen tensors (vector in 1D, matrix
363
- in 2D, cube in 3D), outputs are scalars whether input is real.
364
+ in 2D, cuboid in 3D), outputs are scalars whether input is real.
364
365
\newline Generators: Inputs are noise vectors, outputs are first
365
366
screen tensors.
366
367
*** Image Processing Models
367
368
- Convolutional
368
369
- Dense (MLP)
369
370
370
371
Inputs are first screen tensors, outputs are the constant metadata
371
- vectors also supplied to the sequence predictor.
372
+ vectors and “level has not ended”-bit also supplied to the sequence
373
+ predictor.
372
374
*** Training Loops
373
375
- Checkpointing and resuming training
374
376
- Handling experiments (storing all parameters, logging, ...)
0 commit comments