I often struggle with ML(DL) problems where my accuracy at first attempts is ridiculously low. I was asking myself whether it is a good idea to make a model big enough to overfit then progressively decrease its size to get something that may learn. Thus, follow up the iteration process knowing the range of parameters we have.
It sounds logical for me, can you give me some hints?