#Overfitting

Dennis Alexis Valin Dittrichdavdittrich@fediscience.org
2025-03-17

When Dimensionality Hurts: The Role of #LLM Embedding Compression for Noisy Regression Tasks d.repec.org/n?u=RePEc:arx:pape
"… suggest that the optimal dimensionality is dependent on the signal-to-noise ratio, exposing the necessity of feature compression in high noise environments. The implication of the result is that researchers should consider the #noise of a task when making decisions about the dimensionality of text.

… findings indicate that sentiment and emotion-based representations do not provide inherent advantages over learned latent features, implying that their previous success in similar tasks may be attributed to #regularisation effects rather than intrinsic informativeness."
#ML #autoencoders #Overfitting

2025-01-03

'On the Impact of Hard Adversarial Instances on Overfitting in Adversarial Training', by Chen Liu, Zhichao Huang, Mathieu Salzmann, Tong Zhang, Sabine Süsstrunk.

jmlr.org/papers/v25/22-0950.ht

#adversarial #overfitting #robustness

2024-12-13

"Three married couples. Aren't we just boring now?" joked Mia.
"You're frowning," Ari told Jin.
"Mia's #overfitting. She's too used to seeing threats to relax to realise Tom's mum is fond of her."
"Of course she is, Tom loves Mia. How could his mum not love her too?" #vss365

Daniele de Rigodderigo@hostux.social
2024-09-26

1/

Recent commentary [1]:
escalating concern over the use of the more powerful #chatbots when they are used to go beyond the #knowledge of the human expert who uses them, rather than for simply pre-processing in a controlled way within the domain of human-expert knowledge.

1. ⁠What is often called "hallucination/confabulation” (i.e. severe #extrapolation #uncertainty and #overfitting by the chatbot model) is apparently becoming increasingly realistic with a declining human ability to detect it

2024-08-07

I’m listening to #MITtechReview [PODCAST] Large language models can do jaw-dropping things. But nobody knows exactly why. (7 Aug 2024, 26 min)
pca.st/episode/527cdfed-17ed-4

#edtechSR #MediaLit #AI #lanhuage #learning #DeepLearning #magic #alchemy #OverFitting

AI image created by Wes Fryer with Ideogram:
ideogram.ai/g/FItzSD2BQJiJtpb7

A futuristic, intricate machine with glowing elements is at the center, surrounded by numerous screens displaying cosmic and abstract designs in a dimly lit, high-tech setting.
2024-03-07

"AI in the era of large language models [#LLMs] appears to defy textbook statistics. The most powerful models today are vast, with up to a trillion parameters (the values in a model that get adjusted during training). But statistics says that as models get bigger, they should first improve in performance but then get worse. This is because of something called #overfitting."
technologyreview.com/2024/03/0

pablolarahpablolarah
2023-12-29
Magenta text on light pink:
Eigensolutions
composability as the antidote to overfit
2023-12-23

'Benign Overfitting of Constant-Stepsize SGD for Linear Regression', by Difan Zou, Jingfeng Wu, Vladimir Braverman, Quanquan Gu, Sham M. Kakade.

jmlr.org/papers/v24/21-1297.ht

#overfitting #overparameterized #sgd

2023-11-21

🤖💡 Ever struggled with overfitting in machine learning? It can lead to poor performance and inaccurate predictions. Learn more here 👉 ak-codes.com/overfitting/

Joxean Koret (@matalaz)joxean
2023-09-30

One question for the people: what approach do you use to determine if a decision trees or a random forest approach should work better? Do you simply try both approaches and use whatever seems to work better?

According to what I read, decision trees are more prone to overfitting, while random forest is a more complex approach. Which means little to me 😅

New Submissions to TMLRtmlrsub@sigmoid.social
2023-09-02
Published papers at TMLRtmlrpub@sigmoid.social
2023-08-29

Logistic-Normal Likelihoods for Heteroscedastic Label Noise

Erik Englesson, Amir Mehrpanah, Hossein Azizpour

Action editor: Bo Han.

openreview.net/forum?id=7wA65z

#label #classification #overfitting

New Submissions to TMLRtmlrsub@sigmoid.social
2023-08-07

Label Noise-Robust Learning using a Confidence-Based Sieving Strategy

openreview.net/forum?id=3taIQG

#label #labels #overfitting

Published papers at TMLRtmlrpub@sigmoid.social
2023-08-04

Learning Augmentation Distributions using Transformed Risk Minimization

Evangelos Chatzipantazis, Stefanos Pertigkiozoglou, Kostas Daniilidis, Edgar Dobriban

Action editor: Andriy Mnih.

openreview.net/forum?id=LRYtNj

#augmentation #augmentations #overfitting

wurzelgrumpfjschulze
2023-08-01

🤖𝐂𝐡𝐚𝐭𝐆𝐏𝐓 - 𝐃𝐚𝐬 𝐩𝐞𝐫𝐟𝐞𝐤𝐭𝐞 𝐕𝐞𝐫𝐬𝐩𝐫𝐞𝐜𝐡𝐞𝐧🤖

𝑱𝒆𝒕𝒛𝒕 𝒂𝒍𝒔 𝒆𝑩𝒐𝒐𝒌 𝒆𝒓𝒉𝒂𝒆𝒍𝒕𝒍𝒊𝒄𝒉 𝒃𝒆𝒊 𝑨𝒑𝒑𝒍𝒆𝑩𝒐𝒐𝒌𝒔 𝒖𝒏𝒅 𝑨𝒎𝒂𝒛𝒐𝒏 𝑲𝒊𝒏𝒅𝒍𝒆!

Themen:

⁉️ ⁉️ ⁉️ ⁉️ ⁉️ ⁉️ ⁉️ ⁉️ ⁉️ ⁉️ Souffleuse ⁉️ ⁉️ ⁉️ ⁉️ ⁉️ ⁉️ ⁉️ ⁉️ ⁉️

𝙰𝚞𝚝𝚘𝚛: 𝙹ü𝚛𝚐𝚎𝚗 𝚂𝚌𝚑𝚞𝚕𝚣𝚎
𝚅𝚎𝚛𝚕𝚊𝚐: 𝚒-𝚟𝚊𝚕 - 𝚒𝚗 𝚝𝚑𝚎 "𝚒" 𝚘𝚏 𝚋𝚒𝚣
𝙸𝚂𝙱𝙽: 𝟿𝟽𝟾-𝟹-𝟿𝟷𝟶𝟿𝟷𝟸-𝟶𝟶-𝟽

jschulze.com/projects/ChatGPT/

Published papers at TMLRtmlrpub@sigmoid.social
2023-07-22

Catastrophic overfitting can be induced with discriminative non-robust features

Guillermo Ortiz-Jimenez, Pau de Jorge, Amartya Sanyal et al.

Action editor: Jakub Tomczak.

openreview.net/forum?id=10hCbu

#overfitting #adversarial #robust

zeruchzeruch
2023-07-20

" is leaking into the mainstream in the form of and , but there are serious concerns with this unregulated tech. I'm NOT anti AI, in fact, I believe AI can be of immense benefit to us in the future. But the ethics of AI in its current state MUST be talked about, in order to steer this tech in the right direction."

youtube.com/watch?v=5Viy3Cu3DLk

New Submissions to TMLRtmlrsub@sigmoid.social
2023-06-15

Logistic-Normal Likelihoods for Heteroscedastic Label Noise

openreview.net/forum?id=7wA65z

#label #classification #overfitting

Client Info

Server: https://mastodon.social
Version: 2025.04
Repository: https://github.com/cyevgeniy/lmst