#distributionalSemantics

2024-12-03

My book review of "Lexical variation and change" by Geeraerts et al. is out on linguist list, check it out here:
linguistlist.org/issues/35-340
#linguistics #distributionalSemantics

2024-07-31

In 2013, Mikolov et al. (from Google) published word2vec, a neural network based framework to learn distributed representations of words as dense vectors in continuous space, aka word embeddings.

T. Mikolov et al. (2013). Efficient Estimation of Word Representations in Vector Space. arXiv:1301.3781
arxiv.org/abs/1301.3781

#HistoryOfAI #AI #ise2024 #lecture #distributionalsemantics #wordembeddings #embeddings @sourisnumerique @enorouzi @fizise

Slide from the Information Service Engineering 2024 lecture, Basic Machine Learning 01, a very brief history of AI, Word2Vec. The graphics shows a 2 dimensional carthesian diagrams with word vectors. The text says: Word2Vec neural network based framework to learn distributed representations of words as dense vectors in continuous space (word embeddings) was developed by Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean at Google.
2024-06-08

Besides Wittgenstein, we also quote linguist John Rupert Firth (1890–1960) with "You shall know a word by the company it keeps!" when introducing the principles of distributional semantics as the foundation for word embeddings and large language models.

J.R. Firth (1957), A synopsis of linguistic theory, Studies in linguistic analysis, Blackwell, Oxford: cs.brown.edu/courses/csci2952d

#lecture #llm #nlp #distributionalsemantics @fizise @fiz_karlsruhe @enorouzi @sourisnumerique @shufan #wittgenstein

J.R. Firth as imagined by midjourney, quoting "You shall know a word by the company it keeps" supported by Wittgenstein's quote "The meaning of a word is its use in the language" as for distributional semantics.
2024-06-06

In lecture 05 of our #ise2024 lecture series, we are introducing the concept of distributed semantics and are referring (amongst others) to Ludwig Wittgenstein and his approach to the philosophy of language, and combine it with the idea of word vectors and embeddings.

lecture slides: drive.google.com/file/d/1WcVlk

#wittgenstein #nlp #wordembeddings #distributionalsemantics #lecture @fiz_karlsruhe @fizise @enorouzi @shufan @sourisnumerique #aiart #generativeai

Wittgenstein, as imagined by Midjourney, saying "The meaning of a word is its use in the language". Lecture Information Service Engineering 2024, KIT Karlsruhe
2023-12-10

"You shall know a word by the company it keeps." Distributional Semantics as the basis for self supervised learning as in today's large language models. This is what my colleague @MahsaVafaie and I are going to talk about in this #kg2023 lecture excursion.
OpenHPI video: open.hpi.de/courses/knowledgeg
youtube video: youtube.com/watch?v=EYAY8nmhpd
slides: zenodo.org/records/10185262
@fiz_karlsruhe @fizise @tabea @sashabruns @enorouzi #semantics #distributionalsemantics #llm #llms #cooccurrence #wittgenstein

Slide from the free OpenHPI MOOC Knowledge Graphs - Foundations and Applications, Week 6: Intelligent Applications with Knowledge Graphs and Deep Learning  / Excursion 8: Distributional Semantics and LMs, The slide shows a historical picture of the linguist J. R. Firth as dreamed of by artBot. In a speech bubble he says "You shall know a word by the company it keeps." Below in the slide is the Wittgenstein quote "The meaning of a word is its use in the language". Bibliography: J.R. Firth (1957) A synopsis of linguistic theory, Studies in linguistic analysis, Blackwell, Oxford.
Prompt: “A 1950s photography of J. R. Firth aged 55, with glasses and a white toothbrush mustache, English linguist and leading figure in British linguistics””,  created via ArtBot, ProtoGen, 2022, [CC-BY-4.0], https://tinybots.net/artbot
2023-08-06

Getting ready to leave for #iclc16 in Düsseldorf, here a small teaser for my talk on Tuesday 14:45 on the #affix rivalry between -ity and -ness in #English:
Why to some adjectives take -ity (insular -> insularity), while others take -ness (red -> redness)? Many factors have been considered, I use #distributionalSemantics to explore the role of the adjective's meaning. Mapping the vectors on a two dimensional space with t-SNE, a dimensionality reduction technique, the resulting visualization shows that adjective meaning might indeed be a highly relevant factor. For example, even for adjectives with the same ending -ive, the bases of those taking -ity (e.g. narrativity) and those taking -ness (distinctiveness) fall into two clear clusters.
Looking forward to seeing some of you there :)
#wordformation

Projection of the vectors of the adjective bases ending on -ive into two-dimensional space using the t-SNE dimension reduction technique. Bases of doublets are excluded. The bases taking -ity, in red/yellow, and those taking -ness, in blue, are relatively neatly divided into two clusters.
2023-05-23

Last #NLP chapter of our #ISE2023 lecture last week was on distributional semantics and word embeddings. Of course, Wittgenstein had to be mentioned...

#lecture #distributionalsemantics #wittgenstein #stablediffusionart #creativeAI

@fizise @KIT_Karlsruhe

Cool picture of Wittgenstein as imagined by ArtBot. Wittgenstein is saying "The meaning of a word is its use in the language"
2023-01-18

I somehow just learned about semantic folding
Still trying to learn more about it, but what's really messing with my head is that word embeddings are matrices. Are there any interesting connections to be made between this approach and things like DisCoCat?

#NLP #DistributionalSemantics #DisCoCat

en.wikipedia.org/wiki/Semantic
en.wikipedia.org/wiki/DisCoCat

Client Info

Server: https://mastodon.social
Version: 2025.04
Repository: https://github.com/cyevgeniy/lmst