As we increase the amount of data the performance of the ML model decreases why do DL models take so massive amounts of data?
Hi there,
one reason is that DNNs w/ advanced architectures (like transformers but also architecture w/ convolutional | pooling layers) are designed to perform well on very large datasets and also process highly unstructured data like pictures or videos in a scalable way: basically the more data, the merrier!
Compared to classic ML models, DNNs possess less structure and can learn more complex and abstract relationships given sufficient amount of data.
This thread could be interesting to you since it is about a similar topic:
Best regards
Christian
1 Like
thanks, for the guidance
1 Like
Sure it’s my pleasure. If you have any further questions, pls. do not hesitate to contact me.
Happy learning!
Best
Christian