Skip to content
/ GW-MoE Public

The official implementation of the paper "GW-MoE: Resolving Uncertainty in MoE Router with Global Workspace Theory".

License

Notifications You must be signed in to change notification settings

WaitHZ/GW-MoE

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

GW-MoE

arxiv hf_space

The official implementation of the paper "GW-MoE: Resolving Uncertainty in MoE Router with Global Workspace Theory".

Installation

To install the required environment, execute pip install -r requirements.txt.

Usage

transformers is the code we used to conduct experiments on Switch Transformer.

You can choose a task from ./tasks and run the corresponding bash script to conduct experiments. Before executing the following commands, make sure to set the dataset and model names or paths in the corresponding script.

For text classification:

# Run GWMoE on Switch Transformer for text classification:

bash ./tasks/text-classification/run_glue.sh

For summarization:

# Run GWMoE on Switch Transformer for summarization:

bash ./tasks/summarization/run_summarization.sh

For question-answering:

# Run GWMoE on Switch Transformer for question-answering:

bash ./tasks/question-answering/run_seq2seq_qa.sh

More code is still being organized, and we will update it later.

Citation

If you find our work helpful, please cite our paper:

@misc{wu2024gwmoe,
      title={GW-MoE: Resolving Uncertainty in MoE Router with Global Workspace Theory}, 
      author={Haoze Wu and Zihan Qiu and Zili Wang and Hang Zhao and Jie Fu},
      year={2024},
      eprint={2406.12375},
      archivePrefix={arXiv},
      primaryClass={cs.LG}
}

Acknowledgement

Our codebase is built on Transformers and JetMoE.

License

This source code is released under the MIT license, included here.

About

The official implementation of the paper "GW-MoE: Resolving Uncertainty in MoE Router with Global Workspace Theory".

Resources

License

Stars

Watchers

Forks

Languages