A curated reading list of research in Adaptive Computation, Inference-Time Computation & Mixture of Experts (MoE).
-
Updated
Nov 2, 2024
A curated reading list of research in Adaptive Computation, Inference-Time Computation & Mixture of Experts (MoE).
Exploration into the proposed "Self Reasoning Tokens" by Felipe Bonetto
[Preprint] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models
Yet another random morning idea to be quickly tried and architecture shared if it works; to allow the transformer to pause for any amount of time on any token
(NeurIPS-2019 MicroNet Challenge - 3rd Winner) Open source code for "SIPA: A simple framework for efficient networks"
The ARL Hierarchical MultiScale Framework (ARL-HMS) is a software library for development of multiscale models on heterogeneous high-performance computing systems.
Model implementation for "Adaptive computation as a new mechanism of human attention"
Add a description, image, and links to the adaptive-computation topic page so that developers can more easily learn about it.
To associate your repository with the adaptive-computation topic, visit your repo's landing page and select "manage topics."