Skip to content

Latest commit

 

History

History
41 lines (30 loc) · 1.8 KB

File metadata and controls

41 lines (30 loc) · 1.8 KB

TODOs

Structure reminder

  1. Data acquisition: from the web, a database, a flat file, etc. This includes cleaning the data.
  2. Data exploration: simple exploratory analysis to describe what you got.
  3. Data exploitation: build and train a Machine Learning algorithm based on this data. Any algorithm is considerable, but it has to be motivated.
  4. Evaluation: evaluate the performance, e.g. accuracy, training time, etc., of the chosen model. You define the metrics you care about! If you tried multiple algorithms, please report their performance and try to explain it.

Data acquisition

Data cleaning

  • Convert to greyscale -> original images
    • Should it be rescaled to [0, 255] when generating the orig dataset folder??, e.g. airplane is not whereas arctichare is
  • All images should have the same resolution???
  • images as a dictionary dict()?
  • Split in patches of size 32 x 32 (to be generalized)
  • Center the data
  • Measurement matrix:
    • Gaussian
    • Bernoulli
  • Fixing the seed for reproducible results
    • Check if should be in a file, for now it is defined in the utils.create_measurement_model() function

Data exploration

  • Figure showing an image example, a specific patch and its compressed version (i.e. mixed + blurred)

Data exploitation

Evaluation

Installation

  • Check a docker installation on the NTDS Moodle
  • check_instal.py for manual installation

License