Babak Ehteshami Bejnordi

Research Scientist @ Qualcomm AI Research

Research Projects

Project 1

Cache-MoE

Expo Demo @NeurIPS'24

Efficient Mixture-of-Experts for mobile devices with limited DRAM.

MoE On-device Caching LLM Efficiency
Project 2

Refactor LLM into MoE

Published @NeurIPS'24

Refactorizing LLMs as router-decoupled mixture of experts with system co-design.

MoE Batched-inference Dynamic sparsity Decoupled routing LLM Efficiency
Project 3

LLM-to-SLM

Published @ICML'24: ES-FoMo II

Think Big, Generate Quick: LLM-to-SLM for fast autoregressive decoding.

Hybrid LLM Fast decoding LLM Efficiency LLM to SLM
Project 4

InterroGate for MTL

Published @BMVC'24

Learning to share, specialize, and prune representations for Multi-task Learning.

Multi-task Learning Inference efficiency Gated Networks Channel sparsity
Project 5

Scalarization for MTL

Published @NeurIPS'23

Scalarization for Multi-Task and Multi-Domain Learning at scale.

Population-based Training Scalarization Multi-Task Learning Multi-Domain Learning
Project 6

MSViT

Published @ICCV'23: NIVT

Dynamic mixed-scale tokenization for vision transformers.

Conditional compute Mixed-scale Efficient CV Tokenization
Project 7

Salisa

Published @ECCV'22

Saliency-based input sampling for efficient video object detection.

Efficient Inference VOD Video Object Detection Spatial Transformer Network
Project 8

Single-gated MoE

Published @BMVC'22

Single-gate Mixture of Experts (MoE) with early exiting for convolutional architectures.

MoE Anytime Inference On-device Early-exiting
Project 9

FrameExit

Published @CVPR'21 (Oral)

Conditional Early Exiting for Efficient Video Recognition.

Early Exiting Video Recognition Gating Network Efficient Recognition
Project 10

SkipConv

Published @CVPR'21

Skip-Convolutions for efficient video processing.

Residual Convolutions Efficient Video Processing Skip-Convolution
Project 11

Channel Gating for Continual Learning

Published @CVPR'20 (Oral)

Conditional channel gated networks for task-aware continual learning.

Continual Learning Chanel-Gating Task-aware Dynamic sparsity
Project 12

Channel Gating with Batch-shaping

Published @ICLR'20

Batch-shaping for learning conditional channel gated networks.

Batch-shaping Channel Gating Dynamic sparsity