Search Results for author: Qingyu Zhang

Found 3 papers, 1 papers with code

Base of RoPE Bounds Context Length

no code implementations23 May 2024 Xin Men, Mingyu Xu, Bingning Wang, Qingyu Zhang, Hongyu Lin, Xianpei Han, WeiPeng Chen

We revisit the role of RoPE in LLMs and propose a novel property of long-term decay, we derive that the \textit{base of RoPE bounds context length}: there is an absolute lower bound for the base value to obtain certain context length capability.

Position

ShortGPT: Layers in Large Language Models are More Redundant Than You Expect

no code implementations6 Mar 2024 Xin Men, Mingyu Xu, Qingyu Zhang, Bingning Wang, Hongyu Lin, Yaojie Lu, Xianpei Han, WeiPeng Chen

As Large Language Models (LLMs) continue to advance in performance, their size has escalated significantly, with current LLMs containing billions or even trillions of parameters.

Quantization

MDIA: A Benchmark for Multilingual Dialogue Generation in 46 Languages

1 code implementation27 Aug 2022 Qingyu Zhang, Xiaoyu Shen, Ernie Chang, Jidong Ge, Pengke Chen

In this paper, we present mDIA, the first large-scale multilingual benchmark for dialogue generation across low- to high-resource languages.

Dialogue Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.