Awaker2.5-VL: Stably Scaling MLLMs with Parameter-Efficient Mixture of Experts Paper • 2411.10669 • Published Nov 16, 2024 • 10
CoDeNet: Efficient Deployment of Input-Adaptive Object Detection on Embedded FPGAs Paper • 2006.08357 • Published Jun 12, 2020 • 1
HAO: Hardware-aware neural Architecture Optimization for Efficient Inference Paper • 2104.12766 • Published Apr 26, 2021
SeerAttention: Learning Intrinsic Sparse Attention in Your LLMs Paper • 2410.13276 • Published Oct 17, 2024 • 26
SeerAttention: Learning Intrinsic Sparse Attention in Your LLMs Paper • 2410.13276 • Published Oct 17, 2024 • 26