@scottmiller42 @emovulcan @acute_distress @thomasfuchs
Hey Scott, I have not decided what to think about LLMs and copyrights.
I saw the word Data analyst in your Bio, so I suppose you have some knowledge in #DataScience and #DataEngineering.
I am myself trained in #MachineLearning and a bit in #neuralnetworks but I am far far far far far away from being any close to an expert. I can use them, efficiently sometimes. Sometimes not: there are problems I am still bad at solving with them.
I am still learning the Transformers stuffs and other types of layers used in Deep Learning. That's huge stuff. For me at least.
Could you clear that doubt I have when I read your messages:
when you talk about "MEMORIZING" data
It conflicts with my poor understanding of #machinelearning #neuralnet #transformer
At my level of understanding: memorizing == overfitting
that means that the model cannot extrapolate or even interpolate between values of the training set.
This is something that I have categorized as a real flaw. If any of my models overfits, I consider it broken. I throw it away, it will not perform good on any unexpected input.
Could you explain me how "memorizing" would differ in your vision?
Don't hesitate to simplify things, I am not an expert yet at #transformer and #LLM
In advance thanks. I would not like to be fooled by my apriori.