與「Yi Tay」相符的使用者個人學術檔案

Yi Tay

- 在 google.com 的電子郵件地址已通過驗證 - 被引用 27449 次

Yi Wei Daniel Tay

- 在 ntu.edu.sg 的電子郵件地址已通過驗證 - 被引用 4532 次

Emergent abilities of large language models

J Wei, Y Tay, R Bommasani, C Raffel, B Zoph… - arXiv preprint arXiv …, 2022 - arxiv.org
Scaling up language models has been shown to predictably improve performance and
sample efficiency on a wide range of downstream tasks. This paper instead discusses an …

Quaternion knowledge graph embeddings

S Zhang, Y Tay, L Yao, Q Liu - Advances in neural …, 2019 - proceedings.neurips.cc
In this work, we move beyond the traditional complex-valued representations, introducing
more expressive hypercomplex representations to model entities and relations for knowledge …

Long range arena: A benchmark for efficient transformers

Y Tay, M Dehghani, S Abnar, Y Shen, D Bahri… - arXiv preprint arXiv …, 2020 - arxiv.org
Transformers do not scale very well to long sequence lengths largely because of quadratic
self-attention complexity. In the recent months, a wide spectrum of efficient, fast Transformers …

Deep learning based recommender system: A survey and new perspectives

S Zhang, L Yao, A Sun, Y Tay - ACM computing surveys (CSUR), 2019 - dl.acm.org
With the growing volume of online information, recommender systems have been an effective
strategy to overcome information overload. The utility of recommender systems cannot be …

Sparse sinkhorn attention

Y Tay, D Bahri, L Yang, D Metzler… - … on Machine Learning, 2020 - proceedings.mlr.press
We propose Sparse Sinkhorn Attention, a new efficient and sparse method for learning to
attend. Our method is based on differentiable sorting of internal representations. Concretely, …

[HTML][HTML] Printability region for 3D concrete printing using slump and slump flow test

YWD Tay, Y Qian, MJ Tan - Composites Part B: Engineering, 2019 - Elsevier
Rheological studies are important for successful 3D concrete printing. The main challenge for
successful 3D concrete printing is the complex characteristic the materials should possess. …

Palm: Scaling language modeling with pathways

…, J Maynez, A Rao, P Barnes, Y Tay… - Journal of Machine …, 2023 - jmlr.org
Large language models have been shown to achieve remarkable performance across a variety
of natural language tasks using few-shot learning, which drastically reduces the number …

Scaling instruction-finetuned language models

HW Chung, L Hou, S Longpre, B Zoph, Y Tay… - Journal of Machine …, 2024 - jmlr.org
Finetuning language models on a collection of datasets phrased as instructions has been
shown to improve model performance and generalization to unseen tasks. In this paper we …

Palm 2 technical report

…, M Omernick, K Robinson, S Ruder, Y Tay… - arXiv preprint arXiv …, 2023 - arxiv.org
We introduce PaLM 2, a new state-of-the-art language model that has better multilingual and
reasoning capabilities and is more compute-efficient than its predecessor PaLM. PaLM 2 is …

The flan collection: Designing data and methods for effective instruction tuning

…, T Vu, A Webson, HW Chung, Y Tay… - International …, 2023 - proceedings.mlr.press
We study the design decision of publicly available instruction tuning methods, by reproducing
and breaking down the development of Flan 2022 (Chung et al., 2022). Through careful …