no code implementations • Findings (NAACL) 2022 • Bowen Yang, Cong Han, Yu Li, Lei Zuo, Zhou Yu
In this paper, we propose a simple yet effective architecture comprising a pre-trained language model (PLM) and an item metadata encoder to integrate the recommendation and the dialog generation better.
1 code implementation • 22 Apr 2024 • Yuhong Li, Yingbing Huang, Bowen Yang, Bharat Venkitesh, Acyr Locatelli, Hanchen Ye, Tianle Cai, Patrick Lewis, Deming Chen
Specifically, SnapKV achieves a consistent decoding speed with a 3. 6x increase in generation speed and an 8. 2x enhancement in memory efficiency compared to baseline when processing inputs of 16K tokens.
1 code implementation • 20 Sep 2023 • Nolan Dey, Daria Soboleva, Faisal Al-Khateeb, Bowen Yang, Ribhu Pathria, Hemant Khachane, Shaheer Muhammad, Zhiming, Chen, Robert Myers, Jacob Robert Steeves, Natalia Vassilieva, Marvin Tom, Joel Hestness
BTLM-3B-8K is available under an Apache 2. 0 license on Hugging Face: https://huggingface. co/cerebras/btlm-3b-8k-base.
no code implementations • 19 Sep 2023 • Jie Cheng, Yingbing Chen, Xiaodong Mei, Bowen Yang, Bo Li, Ming Liu
In recent years, imitation-based driving planners have reported considerable success.
1 code implementation • 29 Jun 2023 • Haoqin Tu, Bowen Yang, Xianfeng Zhao
Automatically generating textual content with desired attributes is an ambitious task that people have pursued long.
1 code implementation • 15 Dec 2021 • Bowen Yang, Cong Han, Yu Li, Lei Zuo, Zhou Yu
The encoder learns to map item metadata to embeddings that can reflect the semantic information in the dialog context.
no code implementations • 15 Dec 2021 • Lei Zuo, Kun Qian, Bowen Yang, Zhou Yu
A commonly observed problem of the state-of-the-art natural language technologies, such as Amazon Alexa and Apple Siri, is that their services do not extend to most developing countries' citizens due to language barriers.
no code implementations • 27 Jun 2021 • Bowen Yang, Jing Zhang, Zhenfei Yin, Jing Shao
In practice, given a handful of labeled samples from a new deployment scenario (target domain) and abundant labeled face images in the existing source domain, the FAS system is expected to perform well in the new scenario without sacrificing the performance on the original domain.
no code implementations • 9 Oct 2019 • Bowen Yang, Jian Zhang, Jonathan Li, Christopher Ré, Christopher R. Aberger, Christopher De Sa
Pipeline parallelism (PP) when training neural networks enables larger models to be partitioned spatially, leading to both lower network communication and overall higher hardware utilization.