How to interpret weight distributions of neural ne

2020-07-26 07:19发布

I have designed a 3 layer neural network whose inputs are the concatenated features from a CNN and RNN. The weights learned by network take very small values. What is the reasonable explanation for this? and how to interpret the weight histograms and distributions in Tensorflow? Any good resource for it?

This is the weight distribution of the first hidden layer of a 3 layer neural network visualized using tensorboard. How to interpret this? all the weights are taking up zero value?

This is the weight distribution of the second hidden layer of a 3 layer neural:

1条回答
爱情/是我丢掉的垃圾
2楼-- · 2020-07-26 07:33

how to interpret the weight histograms and distributions in Tensorflow?

Well, you probably didn't realize it, but you have just asked the 1 million dollar question in ML & AI...

Model interpretability is a hyper-active and hyper-hot area of current research (think of holy grail, or something), which has been brought forward lately not least due to the (often tremendous) success of deep learning models in various tasks; these models are currently only black boxes, and we naturally feel uncomfortable about it...

Any good resource for it?

Probably not exactly the kind of resources you were thinking of, and we are well off a SO-appropriate topic here, but since you asked...:

On a more practical level:

These should be enough for starters, and to give you a general idea of the subject about which you asked...

UPDATE (Oct 2018): I have put up a much more detailed list of practical resources in my answer to the question Predictive Analytics - “Why” factor?

查看更多
登录 后发表回答