Skip to yearly menu bar Skip to main content


Poster
in
Workshop: ES-FoMo: Efficient Systems for Foundation Models

SpeedLimit: Neural Architecture Search for Quantized Transformer Models

Luke Bailey · Yuji Chai · Yunho Jin · Glenn Ko · Matthew Karle


Abstract:

While prevailing research in the field of transformer models has primarily focused on enhancing performance metrics such as accuracyand perplexity, practical applications in industry often necessitate a rigorous consideration of inference latency constraints. Addressing this challenge, we introduce SpeedLimit, a novel Neural Architecture Search (NAS) technique that optimizes accuracy whilst adhering to an upper-bound latency constraint. Our method incorporates 8-bit integer quantization in the search process to outperform the current state-of-the-art technique. Our results underline the feasibility and efficacy of seeking an optimal balance between performance and latency, providing new avenues for deploying state-of-the-art transformer models in latency-sensitive environments.

Chat is not available.