Skip to content

8-bit quantized #4

@jerinphilip

Description

@jerinphilip

Measure 8-bit quantized doing any better? I expect it to be lossy, but definitely faster? 32 downto 8 is a 4x speedup at minimum?
If intel can swap out for intgemm as well, eliminating the dependency situation.

browsermt/marian-dev#50

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions