Search Results for author: WeiPeng Chen

Found 8 papers, 3 papers with code

Exploring Context Window of Large Language Models via Decomposed Positional Vectors

no code implementations28 May 2024 Zican Dong, Junyi Li, Xin Men, Wayne Xin Zhao, Bingbing Wang, Zhen Tian, WeiPeng Chen, Ji-Rong Wen

Based on our findings, we design two training-free context window extension methods, positional vector replacement and attention window extension.

Base of RoPE Bounds Context Length

no code implementations23 May 2024 Xin Men, Mingyu Xu, Bingning Wang, Qingyu Zhang, Hongyu Lin, Xianpei Han, WeiPeng Chen

We revisit the role of RoPE in LLMs and propose a novel property of long-term decay, we derive that the \textit{base of RoPE bounds context length}: there is an absolute lower bound for the base value to obtain certain context length capability.

Position

Checkpoint Merging via Bayesian Optimization in LLM Pretraining

no code implementations28 Mar 2024 Deyuan Liu, Zecheng Wang, Bingning Wang, WeiPeng Chen, Chunshan Li, Zhiying Tu, Dianhui Chu, Bo Li, Dianbo Sui

The rapid proliferation of large language models (LLMs) such as GPT-4 and Gemini underscores the intense demand for resources during their training processes, posing significant challenges due to substantial computational and environmental costs.

Bayesian Optimization

ShortGPT: Layers in Large Language Models are More Redundant Than You Expect

no code implementations6 Mar 2024 Xin Men, Mingyu Xu, Qingyu Zhang, Bingning Wang, Hongyu Lin, Yaojie Lu, Xianpei Han, WeiPeng Chen

As Large Language Models (LLMs) continue to advance in performance, their size has escalated significantly, with current LLMs containing billions or even trillions of parameters.

Quantization

ComQA:Compositional Question Answering via Hierarchical Graph Neural Networks

1 code implementation16 Jan 2021 Bingning Wang, Ting Yao, WeiPeng Chen, Jingfang Xu, Xiaochuan Wang

In compositional question answering, the systems should assemble several supporting evidence from the document to generate the final answer, which is more difficult than sentence-level or phrase-level QA.

Answer Selection Machine Reading Comprehension +2

Cannot find the paper you are looking for? You can Submit a new open access paper.