μ μ μμ Labled dataμ λ§μ μμ Unlabeled dataλ‘ νμ΅μν€λ technique
- Gaussian Mixture Model
- Deep Generative Model
- Label Propagation
Pass massages between pairs of nodes and agglomerate
- Notation
G = (A, X)
: GraphA
: Adjacency matrixX
: Node feature matrix
- Task
- Predict node label of unlabeled nodes
- Self-Training
- Co-Training
- Train the model with labeled data
- Predict the unlabeled data using the trained model
- Add most confident (unlabeled) sample to the labeled data
- Repeat 1-3
e.g., Pi-Model, Mean Teacher, ...
- Train the model with labeled data
- Predict the unlabeled data using the trained model
- Add noise to the unlabeled data (i.e., data augmentation)
- Train the model with the augmented unlabeled data where the ground truth is the predicted labels from 2.
- Repeat 1-4
- Expense of producing a new dataset for each task
- Take advantage of vast of unlabeled data on the internet
- Successful story of supervised learning comes from the utility of pre-trained models for various downstream tasks
- e.g., detection, segmentation, ...
- Learn equally good (if not better) features without supervision
- Generalize better potentially because you learn more about the world
- Input: Corrupted data (Noised data)
- Output: Original data
- Input: Shuffled image patches
- Output: Relative position of image patches
- Input: Shuffled image patches
- Output: Original image patches
- Input: Rotated image
- Output: Rotation angle
- Adjust a model from the source domain knowledge to a different but related target domain
- Unsupservised DA
- No labeled in the target domain
- Domain-invariant feature learning
- Pixel-level DA using GAN
Knowledge distillation is a process of distilling or transferring the knowledge from a large model(s) to a lighter, easier-to-deploy single model, without significant loss in performance.
- μΈκ³΅μ§λ₯ μμ© (ICE4104), μΈνλνκ΅ μ 보ν΅μ 곡νκ³Ό νμ±μ κ΅μλ