NLP

大语言模型能否具备空间智能?
大语言模型能否具备空间智能?

(这篇博文用英文攥写并翻译自ChatGPT) 最近,我们学校庆祝了计算机科学与人工智能专业成立60周年。为了纪念这一时刻,组织者邀请了Fernando Pereira发表关于语言形式与意义之间联系的演讲。多年来,这个主题一直吸引着语言学家、计算机科学家和认知研究人员的关注。 在演讲中,Pereira提出了一个发人深省的例子。他提问:“Bob坐在Alice的右边,Jack在Bob的左边。如果每个人都将自己的名牌传给右边的人,最后谁会拿到哪个名牌? 令人惊讶的是,当这个问题被测试在一个大语言模型(LLM)上时,它未能给出正确答案(我自己还没试过,但这并不重要)。这引发了我的好奇:大语言模型能否获得空间智能? 我想到的一个潜在解决方案是将文本到图像的组件与文本编码器结合起来。通过生成图像,我们或许能够保留语言编码器遗漏的空间信息。 为了获得更多见解,我联系了哥伦比亚大学一位专攻计算机视觉的朋友。然而,他表达了怀疑态度。他指出,当前的图像生成模型,如稳定扩散、DALLE和Imagen,在文本理解上严重依赖于CLIP。因此,如果大语言模型在理解空间关系上遇到困难,那么图像生成也可能会面临同样的挑战。 尽管我们都认识到将问题转换为代码或逻辑推理可能会解决这一问题,但这偏离了最初的目标:赋予大语言模型空间智能。而且,随着空间场景变得越来越复杂,创建逻辑表示变得愈发困难。 暂时先把这个想法留在这里。随着我深入研究这一课题并收集更多见解,我也许会更新这篇文章。

10月 13, 2023

BERT Is Not The Count: Learning to Match Mathematical Statements with Proofs

Citation @inproceedings{li-etal-2023-bert, title = "{BERT} Is Not The Count: Learning to Match Mathematical Statements with Proofs", author = "Li, Weixian Waylon and Ziser, Yftah and Coavoux, Maximin and Cohen, Shay B.", editor = "Vlachos, Andreas and Augenstein, Isabelle", booktitle = "Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics", month = may, year = "2023", address = "Dubrovnik, Croatia", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2023.eacl-main.260", doi = "10.18653/v1/2023.eacl-main.260", pages = "3581--3593", abstract = "We introduce a task consisting in matching a proof to a given mathematical statement. The task fits well within current research on Mathematical Information Retrieval and, more generally, mathematical article analysis (Mathematical Sciences, 2014). We present a dataset for the task (the MATcH dataset) consisting of over 180k statement-proof pairs extracted from modern mathematical research articles. We find this dataset highly representative of our task, as it consists of relatively new findings useful to mathematicians. We propose a bilinear similarity model and two decoding methods to match statements to proofs effectively. While the first decoding method matches a proof to a statement without being aware of other statements or proofs, the second method treats the task as a global matching problem. Through a symbol replacement procedure, we analyze the {``}insights{''} that pre-trained language models have in such mathematical article analysis and show that while these models perform well on this task with the best performing mean reciprocal rank of 73.7, they follow a relatively shallow symbolic analysis and matching to achieve that performance.", }

5月 1, 2023

对比学习简介
对比学习简介

(这篇博文用英文攥写并翻译自ChatGPT) 有关阅读材料: The Beginner’s Guide to Contrastive Learning SimCSE: Simple Contrastive Learning of Sentence Embeddings A Simple Framework for Contrastive Learning of Visual Representations Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks 背景 对比学习旨在通过将语义上接近的邻居聚集在一起并将非邻居推开来学习有效的表示。最初,对比学习应用于计算机视觉任务。如下图所示,我们期望模型学习具有相同标签的两个图像之间的关联性,以及具有不同标签的图像对之间的差异。

8月 8, 2022

Real-Time Speech Recognition Using Python
Real-Time Speech Recognition Using Python

A step-by-step guidance to create a simple real-time speech recogniser

3月 5, 2020