模型压缩95%:Lite Transformer,MIT韩松等人 Lite Transformer with Long-Short Range Attention Zhanghao Wu, Zhijian Liu, Ji Lin, Yujun Lin, Song Han 论文地址:https://arxiv.org/abs/2004.11886v1 GitHub 地址:https://github.com/mit-han-lab/lite-transformer 摘要 Transformer在自…
SP的提出时间比CoSaMP提出时间稍晚一些,但和压缩采样匹配追踪(CoSaMP)的方法几乎是一样的.SP与CoSaMP主要区别在于“In each iteration, in the SP algorithm, only K new candidates are added, while theCoSAMP algorithm adds 2K vectors.”,即SP每次选择K个原子,而CoSaMP则选择2K个原子:这样带来的好处是“This makes the SP algorithm c…