Skip to content
Zhengyuan Zhu
Go back

Introduction to Meta-Learning

Two problems we confront

So meta-learning is the solution to the two questions above. And we try to define it as “learning how to learn”. Our dream is:

Some basic concepts

Few-shot Learning

In deep learning, we use regularization to make sure we are not overfitting out model with a small dataset, but we are overfitting our task. Therefore what we learned cannot be generalized to other tasks.

We often get stuck when test samples that are not common in dataset.

In one-shot-learning, we will only provide one training sample per category. There is an example:

In this one-shot learning, we often train a RNN to learn the training data and labels. When we represent with a test input, we should predict its label correctly.

In meta-testing, we provide many datasets again with classes that never trained before. Once we have learned from hundred tasks, we should discover the general pattern in classifying objects.

Recurrent Models

Memory-Augmented Neural Networks

One of the meta-learning methods using an external memory network with RNN. Note that in supervised learning, we provide both input and label in the same time step $t$. However, in this model, the label is not provided untild the next time step $t+1$(shown below).

When updating the model, instead of updating the model immediately, we wait until a batch of tasks is completed. We later merge all we learned from these tasks for a single update.

Reference


Share this post on:

Previous Post
Artificial General Intelligence (AGI)
Next Post
Capsule Network
Jack the orange tabby cat
I'm Jack 🧡
Luna the tuxedo cat
I'm Luna! 🖤