suppose i have 10 individual observations each of size (125,59). i want to group these 10 observations based on their 2d feature matrices ((125,59)).Is this possible without flattening every observation to 125*59 1D matrix ? I cant even implement PCA or LDA for feature extraction because the data is highly variant. Please note that i am trying to implement clustering through self organizing maps or neural networks. Deep learning and neural networks are completely related to the question asked.
相关问题
- How to use Reshape keras layer with two None dimen
- How to conditionally scale values in Keras Lambda
- neural network does not learn (loss stays the same
- Trying to understand Pytorch's implementation
- Convolutional Neural Network seems to be randomly
相关文章
- how to flatten input in `nn.Sequential` in Pytorch
- How to downgrade to cuda 10.0 in arch linux?
- How to use cross_val_score with random_state
- How to pick the T1 and T2 threshold values for Can
- Looping through training data in Neural Networks B
- Why does this Keras model require over 6GB of memo
- How to measure overfitting when train and validati
- McNemar's test in Python and comparison of cla
Of course it is.
Define an appropriate distance measure.
Then compute the 10x10 distance matrix, and run hierarchical clustering.