Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> Some interesting research has been done on the "loss landscape" of these giant neural network models, and my understanding is that they are messy and complicated.

Do you have any recommended reading for this? It sounds like a super interesting area of research.



This is the one example I had in mind because of all the pretty pictures: https://arxiv.org/abs/1712.09913

Hao Li, Zheng Xu, Gavin Taylor, Christoph Studer, Tom Goldstein. 2018. "Visualizing the Loss Landscape of Neural Nets".




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: