大地小神
你们都是大傻瓜, 我是天下大赢家
-
-
MSE和CE不同
dummy What's the difference between cross-entropy and mean-squared-error loss functions?
These loss functions have different derivatives and different purposes. From a probabilistic point of view, the cross-entropy arises as the natural cost function to use if you have a sigmoid or softmax nonlinearity in the output layer of your network, and you want to maximize the likelihood of classifying the input data correctly.
If instead you assume the target is continuous and normally distributed, and you maximize the likelihood of the output of the net under these assumptions, you get the MSE (combined with a linear output layer).
For classification, cross-entropy tends to be more suitable than MSE -- the underlying assumptions just make more sense for this setting. That said, you can train a classifier with the MSE loss and it will probably work fine (although it does not play very nicely with the sigmoid/softmax nonlinearities, a linear output layer would be a better choice in that case). For regression problems, you would almost always use the MSE.
Another alternative for classification is to use a margin loss, which basically amounts to putting a (linear) SVM on top of your network.
-
(原创)交叉熵实例
-
数学符号表
-
(原创)交叉熵损失函数
-
(原创)本地规则引擎设计
-
(原创)用数据模拟卷积-初探
dummy 网上很多写卷积的文档都是用"很简单"表示卷积so easy, 但草民认真读了他们的文章后, 反而对 卷积更不理解, 即使到现在还是不能理解, 这篇文章只是从数据上分析卷积过程, 很多疑点还需继 续探究
阅读全文… -
Jupyter支持动态图
dummy 1 参考
http://louistiao.me/posts/notebooks/embedding-matplotlib-animations-in-jupyter-notebooks/
http://louistiao.me/posts/notebooks/embedding-matplotlib-animations-in-jupyter-as-interactive-javascript-widgets/
阅读全文… -
A Student's Guide to Vectors and Tensors
-
(原创)简单感知机-初探