Skip to content

Latest commit

 

History

History
27 lines (12 loc) · 969 Bytes

README.md

File metadata and controls

27 lines (12 loc) · 969 Bytes

The Sparsely Gated Mixture of Experts Layer for PyTorch

source: https://techburst.io/outrageously-large-neural-network-gated-mixture-of-experts-billions-of-parameter-same-d3e901f2fe05

This repository contains the PyTorch re-implementation of the MoE layer described in the paper Outrageously Large Neural Networks for PyTorch.

Requirements

This example was tested using torch v1.0.0 and Python v3.6.1 on CPU.

To install the requirements run:

pip install -r requirements.txt

Example

The file test.py contains an example illustrating how to train and evaluate the MoE layer with dummy inputs and targets. To run the example:

python test.py

Acknowledgements

The code is based on the TensorFlow implementation that can be found here.