HI Sir,
@thearkamitra
@arosacastillo
@AmmarMohanna
@XpRienzo
@reinoudbosch
@chrismoroney39
@paulinpaloalto
In the lecture video “state of computer vision”, we had doubts in the below following statement. We are unable to understand many statements told by andrew ng, can you please help to clarify ?
Statement 1:
So, if you look across a broad spectrum of machine learning problems, you see on average that when you have a lot of data you tend to find people get in way with using simpler algorithms as well as less hand-engineering. So, there’s just less needing to carefully design features for the problem, but instead you can have a giant neural network, even a simpler architecture, and have a neural network.
Just learn whether we want to learn we have a lot of data.
Doubt 1: instead you can have a giant neural network, even a simpler architecture. what does it meaning the statement cannot understand. simpler architecture but giant neural network Ah!..what does it mean sir ?
Doubt 1b: Why we need simpler architecture when we have lots of more data ? Why we need complex architecture when we have little data ?
Doubt 2: What is the reason behind less hand engineering things when lots of data and more hand engineering things when small data ? why so sir like that ? cannot understand the reason …
Statement 2: But there’s still a lot of hand-engineering of network architectures and computer vision. Which is why you see very complicated hyper frantic choices in computer vision, are more complex than you do in a lot of other disciplines.
Doubt 3 : what does it means lot of hand engineering NN architecture become leads to complicated hyperparameter choices? can u give some example to understand ? what is hand engineering NN architecture ? and How it becomes lead to complicate choice of hyperparamter ?
Doubt 4: What does it means standardized benchmark datasets ? what does it mean the context of benchmark ?
Statement 3 : But you also see in the papers people do things that allow you to do well on a benchmark, but that you wouldn’t really use in a production or a system that you deploy in an actual application.
Regarding statement 3, Preferred network in benchmark not used in production system. Are the these due to Reason like computational budget. Can u please let us know what covers under computational budget ? And other reasons are lot more memory , slow down running time right ? can u please confirm others reasons are correct ?
Statement 4: Train several NN independently and average their output? Here what average their output y hats means accuracy of the result?
General doubt :
As per the lecture, Ensembling technique actually tends to more focus on testing time rather than training & cross validation because some things we observed from the lecture. Is it true ensembling actually more focus on test time rather than training time & Cross Validation ? because im saying according to the below points observed from the lecture.
-
Train several Neural networks independently and average their results. This is the statement from the lecture. Does it means train NN at against training dataset & CRoss validation sets or at test time?
-
But because ensembling means that to test on each image, you might need to run an image through anywhere from say 3 to 15 different networks quite typical. This is the statement from the lecture. Does it means train 3 to 15 NN at test time ? Not right ?
3.Why need of run classifier for multiple crops at test time ? We can do data augmentation at training time right. Thats why we asked is ensembling tends to focus on test time more ? why it should be like that ?
can u please help to understand please ?