no code implementations • 16 Nov 2023 • Qirui Tang, Wenkang Jiang, Yihua Du, Lei Lin
In social media networks, users produce a large amount of text content anytime, providing researchers with a valuable approach to digging for personality-related information.
no code implementations • 14 Nov 2023 • Lei Lin, Jiayi Fu, Pengli Liu, Qingyang Li, Yan Gong, Junchen Wan, Fuzheng Zhang, Zhongyuan Wang, Di Zhang, Kun Gai
Although chain-of-thought (CoT) prompting combined with language models has achieved encouraging results on complex reasoning tasks, the naive greedy decoding used in CoT prompting usually causes the repetitiveness and local optimality.
1 code implementation • 30 Oct 2023 • Tianwen Wei, Liang Zhao, Lichang Zhang, Bo Zhu, Lijie Wang, Haihua Yang, Biye Li, Cheng Cheng, Weiwei Lü, Rui Hu, Chenxia Li, Liu Yang, Xilin Luo, Xuejie Wu, Lunan Liu, Wenjun Cheng, Peng Cheng, Jianhao Zhang, XiaoYu Zhang, Lei Lin, Xiaokun Wang, Yutuan Ma, Chuanhai Dong, Yanqi Sun, Yifu Chen, Yongyi Peng, Xiaojuan Liang, Shuicheng Yan, Han Fang, Yahui Zhou
In this technical report, we present Skywork-13B, a family of large language models (LLMs) trained on a corpus of over 3. 2 trillion tokens drawn from both English and Chinese texts.
1 code implementation • 25 Oct 2023 • Liu Yang, Haihua Yang, Wenjun Cheng, Lei Lin, Chenxia Li, Yifu Chen, Lunan Liu, Jianfei Pan, Tianwen Wei, Biye Li, Liang Zhao, Lijie Wang, Bo Zhu, Guoliang Li, Xuejie Wu, Xilin Luo, Rui Hu
Large language models (LLMs) have shown great potential to solve varieties of natural language processing (NLP) tasks, including mathematical reasoning.
no code implementations • 11 Oct 2023 • Jiayi Fu, Lei Lin, Xiaoyang Gao, Pengli Liu, Zhengzong Chen, Zhirui Yang, ShengNan Zhang, Xue Zheng, Yan Li, Yuliang Liu, Xucheng Ye, Yiqiao Liao, Chao Liao, Bin Chen, Chengru Song, Junchen Wan, Zijia Lin, Fuzheng Zhang, Zhongyuan Wang, Di Zhang, Kun Gai
Recent advancements in large language models (LLMs) have demonstrated remarkable abilities in handling a variety of natural language processing (NLP) downstream tasks, even on mathematical tasks requiring multi-step reasoning.
Ranked #88 on Arithmetic Reasoning on GSM8K (using extra training data)
no code implementations • 20 Jul 2023 • Yafang Zheng, Lei Lin, Shuangtao Li, Yuxuan Yuan, Zhaohong Lai, Shan Liu, Biao Fu, Yidong Chen, Xiaodong Shi
Inspired by this, we propose LRF, a novel \textbf{L}ayer-wise \textbf{R}epresentation \textbf{F}usion framework for CG, which learns to fuse previous layers' information back into the encoding and decoding process effectively through introducing a \emph{fuse-attention module} at each encoder and decoder layer.
no code implementations • 20 May 2023 • Lei Lin, Shuangtao Li, Yafang Zheng, Biao Fu, Shan Liu, Yidong Chen, Xiaodong Shi
There is mounting evidence that one of the reasons hindering CG is the representation of the encoder uppermost layer is entangled, i. e., the syntactic and semantic representations of sequences are entangled.
no code implementations • 21 Mar 2023 • Lei Lin, Shuangtao Li, Xiaodong Shi
Simultaneous machine translation, which aims at a real-time translation, is useful in many live scenarios but very challenging due to the trade-off between accuracy and latency.
no code implementations • 7 Oct 2022 • Boyang Zhang, Suping Wu, Hu Cao, Kehua Ma, Pan Li, Lei Lin
Different from them, our STR aims to learn accurate and natural motion sequences in an unconstrained environment through temporal and spatial tendency and to fully excavate the spatio-temporal features of existing video data.
Ranked #55 on 3D Human Pose Estimation on MPI-INF-3DHP
1 code implementation • 22 Nov 2021 • Lei Lin, Weizi Li, Lei Zhu
For instance, our model reduces MAE by 25. 3%, RMSE by 29. 2%, and MAPE by 20. 2%, compared to the state-of-the-art Diffusion Convolutional Recurrent Neural Network (DCRNN) model using the hourly dataset.
no code implementations • 6 Nov 2021 • Zhongkai Shangguan, Zihe Zheng, Lei Lin
Nowadays social media platforms such as Twitter provide a great opportunity to understand public opinion of climate change compared to traditional survey methods.
1 code implementation • 21 Aug 2021 • Han Sun, Lei Lin, Ningzhong Liu, Huiyu Zhou
In this paper, we propose a Robust Ensembling Network (REN) for UDA, which applies a robust time ensembling teacher network to learn global information for domain transfer.
no code implementations • SEMEVAL 2021 • Genyu Zhang, Yu Su, Changhong He, Lei Lin, Chengjie Sun, Lili Shan
This paper describes the winning system in the End-to-end Pipeline phase for the NLPContributionGraph task.
no code implementations • ACL 2021 • Yadong Xi, Xiaoxi Mao, Le Li, Lei Lin, Yanjiang Chen, Shuhan Yang, Xuhan Chen, Kailun Tao, Zhi Li, Gongzheng li, Lin Jiang, Siyan Liu, Zeng Zhao, Minlie Huang, Changjie Fan, Zhipeng Hu
Equipped with GPT-2 and the latest GPT-3, AI Dungeon has been seen as a famous example of the powerful text generation capabilities of large-scale pre-trained language models, and a possibility for future games.
no code implementations • 24 Jul 2021 • Lei Lin
However, we argue that there are still gaps between the predictor and the estimator in both data quality and training objectives, which preclude QE models from benefiting from a large number of parallel corpora more directly.
no code implementations • 3 Apr 2021 • Zhongkai Shangguan, Lei Lin, Wencheng Wu, Beilei Xu
In order to validate the benefits of NPBO, we compare NPBO with four benchmark approaches on a power system parameter optimization problem and a series of seven benchmark Bayesian optimization problems.
1 code implementation • CVPR 2020 • Trevor W. Richardson, Wencheng Wu, Lei Lin, Beilei Xu, Edgar A. Bernal
We consider the topic of data imputation, a foundational task in machine learning that addresses issues with missing data.
no code implementations • 19 Apr 2019 • Tao Li, Lei Lin
", "how private is it?
no code implementations • 28 Mar 2019 • Lei Lin, Beilei Xu, Wencheng Wu, Trevor Richardson, Edgar A. Bernal
Myotonia, which refers to delayed muscle relaxation after contraction, is the main symptom of myotonic dystrophy patients.
1 code implementation • 30 Jul 2018 • Tao Li, Lei Lin, Minsoo Choi, Kaiming Fu, Siyuan Gong, Jian Wang
With one billion monthly viewers, and millions of users discussing and sharing opinions, comments below YouTube videos are rich sources of data for opinion mining and sentiment analysis.
no code implementations • SEMEVAL 2018 • Wenjie Liu, Chengjie Sun, Lei Lin, Bingquan Liu
Semantic Evaluation (SemEval) 2018 Task 12 {``}The Argument Reasoning Comprehension{''} committed to research natural language reasoning.
no code implementations • 13 Dec 2017 • Lei Lin, Zhengbing He, Srinivas Peeta
Two architectures of the GCNN-DDGF model are explored; GCNNreg-DDGF is a regular GCNN-DDGF model which contains the convolution and feedforward blocks, and GCNNrec-DDGF additionally contains a recurrent block from the Long Short-term Memory neural network architecture to capture temporal dependencies in the bike-sharing demand series.
no code implementations • 13 Nov 2017 • Lei Lin, John Handley, Adel Sadek
Short-term traffic volume prediction models have been extensively studied in the past few decades.
no code implementations • 31 Oct 2017 • Zhenhua Zhang, Lei Lin
It can identify the abnormal spatial-temporal patterns for both passenger vehicles and trucks at three bridges connecting US and Canada.
no code implementations • SEMEVAL 2017 • Wenjie Liu, Chengjie Sun, Lei Lin, Bingquan Liu
Semantic Textual Similarity (STS) devotes to measuring the degree of equivalence in the underlying semantic of the sentence pair.
no code implementations • 20 Jan 2017 • Lei Lin, Qian Wang, Adel W. Sadek
Both the proposed method based on the FP tree algorithm, as well as the widely utilized, random forest method, were then used to identify the important variables or the Virginia dataset.
2 code implementations • 30 May 2016 • Yang Liu, Chengjie Sun, Lei Lin, Xiaolong Wang
In our approach, the encoding of sentence is a two-stage process.
Ranked #73 on Natural Language Inference on SNLI
no code implementations • 18 Apr 2014 • Yaming Sun, Lei Lin, Duyu Tang, Nan Yang, Zhenzhou Ji, Xiaolong Wang
We present a method to leverage radical for learning Chinese character embedding.