Any-Shot Learning From Multimodal Observations (ALMO)
Any-Shot Learning From Multimodal Observations (ALMO)
Blog Article
princess polly dresses long sleeve In this paper, we propose a framework (ALMO) for any-shot learning from multi-modal observations.Using training data containing both objects (inputs) and class attributes (side information) from multiple modalities, ALMO embeds the high-dimensional data into a common stochastic latent space using modality-specific encoders.Subsequently, a non-parametric classifier is trained to predict the class labels of the objects.We perform probabilistic data fusion to combine the modalities in the stochastic latent space and learn class earthbath facial wipes conditional distributions for improved generalization and scalability.
We formulate ALMO for both few-shot and zero-shot classification tasks, demonstrating significant improvement in recognition performance on the Omniglot and CUB-200 datasets as compared to state-of-the-art baselines.