You’ve raised two important points. Let’s address them one by one:
Even though your primary interest lies in the latent space representations for clustering, the encoding phase cannot be bypassed. In an autoencoder architecture, the encoder and decoder work together during training to create meaningful representations in the latent space. Here’s why the encoding phase is essential:
-
Learning Compressed Representations: The encoder learns to compress the input data into a lower-dimensional latent space. This process forces the network to capture the most important features of the data.
-
Ensuring Quality of Representations: The decoder’s task of reconstructing the original data from the latent representation ensures that this representation is meaningful. Without the reconstruction requirement, the encoder might learn a trivial or ineffective compression.
-
End-to-End Training: The encoder and decoder are trained together. This joint training ensures that the latent space representations are optimized for capturing the relevant aspects of the input data.
Evaluating unsupervised learning models, especially in the context of clustering, can be challenging. However, there are several strategies you can employ:
-
Using Labels for Evaluation (If Available):
- External Validation: If you have labels available for a subset of your data (even if they weren’t used during training), you can use them for external validation. Metrics like Adjusted Rand Index, Normalized Mutual Information, or Purity can help assess how well your clusters correspond to labeled categories.
- Semi-supervised Learning: Labels can also be used in a semi-supervised learning setting where you initially train your model in an unsupervised manner and then fine-tune or validate using the available labels.
-
Without Labels:
- Intrinsic Metrics: Use intrinsic metrics like Silhouette Score, Davies-Bouldin Index, or Calinski-Harabasz Index to evaluate the quality of the clusters based on the data alone, without any external labels.
- Qualitative Analysis: Sometimes, especially with complex data like mocap, qualitative evaluation can be valuable. This might involve visually inspecting clustered sequences or having domain experts assess the coherence and relevance of the clusters.
- Consistency Checks: Check for consistency across different runs of the algorithm or when varying parameters (like the number of clusters). Consistent results can indicate stable clustering.
-
Reconstruction Quality:
- While your primary goal isn’t reconstruction, assessing the reconstruction quality can still provide insights into whether the latent space captures enough information about the input data.
-
Domain-Specific Evaluation: Depending on your specific application or the nature of your mocap data, there might be domain-specific evaluation methods that could be relevant.
In summary, while the encoding phase is an integral part of the training process for an LSTM autoencoder, you don’t necessarily need to use the decoding part in your clustering application. For evaluation, leveraging any available labels can provide a more straightforward validation method, but in their absence, intrinsic metrics and qualitative analyses are key tools.