Search Results for author: Neha Prakriya

Found 1 papers, 1 papers with code

HMT: Hierarchical Memory Transformer for Long Context Language Processing

1 code implementation9 May 2024 Zifan He, Zongyue Qin, Neha Prakriya, Yizhou Sun, Jason Cong

With an additional 0. 5% - 2% of parameters, HMT can easily plug in and augment future LLMs to handle long context effectively.

Language Modelling Memorization +1

Cannot find the paper you are looking for? You can Submit a new open access paper.