OSLSM(One-Shot Learning for Semantic Segmentation)[1] firstly proposed two-branch approach to one-shot semantic segmentation. Conditioning branch trains a network
to get parameter
$\theta$, and Segmentaion branchoutputs the final mask
based on parameter $\theta$. There are some details of reading and implementing it.
Contents
Paper & Code & note
Paper: One-Shot Learning for Semantic Segmentation(BMVC 2017 paper)
Code: Caffe
Note: Mendeley
Paper
Abstract
- They extend
low-shot methods
to supportsemantic segmentation
.- They train a network that
produces parameters
for a FCN.- They use this FCN to perform dense
pixel-level prediction
on a test image for thenew semantic class
.- It outperforms the state-of-the-art method on the
PASCAL VOC 2012
dataset.
Problem Description
A simple approach to performing one-shot semantic image segmentation is to fine-tune a pre-trained segmentation network on the labeled image.
- This approach
is prone to over-fitting
due to the millions of parameters being updated. - The fine tuning approach to one-shot learning, which
may require many iterations of SGD
to learn parameters for the segmentation network. - Besides, thousands of dense features are computed from a single image and one-shot methods
do not scale well to this many features
.
Problem Solution
Conceptual Understanding
Core Conception
Experiments
Code
[Updating]
Note
- It takes inspiration from few-shot learning and firstly proposes a novel two-branched approach to one-shot semantic segmentation.
References
[1] Shaban A, Bansal S, Liu Z, et al. One-shot learning for semantic segmentation[J]. arXiv preprint arXiv:1709.03410, 2017.
[2] OSLSM. https://github.com/lzzcd001/OSLSM.