Skip to content

Proposal: Add standardized eye-tracking benchmark evaluation endpoint #79

@Sahil-aka

Description

@Sahil-aka

Hi,

I implemented a structured benchmarking module for evaluating eye-tracking accuracy and precision, aligned with the project’s validation goals.

Key additions:

  • Accuracy metrics (mean, median, p95) in pixels and degrees
  • Scientifically grounded per-target RMS precision metric
  • Data quality reporting
  • Per-target accuracy breakdown
  • New API endpoint: POST /api/session/benchmark
  • Input validation and NumPy serialization handling

This enables reproducible and standardized benchmarking across devices and setups.

Before opening a PR, I’d like feedback on:

  • Metric definitions
  • API structure
  • Naming conventions
  • Integration approach

Looking forward to your thoughts.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions