Jul 23, 2017 · 1 min read
The findings in the video above don’t surprise me even the slightest. When you train a network on N classes and then approach it with something completely different, of course it will predict crap. It’s DL chapter 1 that the confidence of an output has nothing to do with what we call probability. The other mentioned paper from 2013 was interesting though.
