title |
---|
segmentation |
Now it's time to have a dedicated page for segmentation.
From J. Malik lab, so should read it. Use features at all scales for segmentation. many other tricks, too.
The author thought about possibility of defining a classifier at each pixel, because features do not have spatial information. But to define a location specific classifier for each of 50x50 pixel is too many parameters. So they define a sparser set of classifiers and interpolate to get all pixels. That remind me the U-net paper that just use last 1x1 convolutional layer as classifier, and that amounts to have a classifier at each location?
When using features at all scales of CNN, we get a feature matrix of
Up-sampling the top layers so the feature map has same resolution with previous layers.
This is a well-written paper. Very readable.
Yet another segmentation paper.
Weakly- and Semi-Supervised Learning of a Deep Convolutional Network for Semantic Image Segmentation
This one also cites the Fully convnet paper. Particular interesting because it's weakly supervised.
Image segmentation, too. But have a unified learning framework including the feature learning and segmentation.
Mapping Stacked Decision Forests to Deep and Sparse Convolutional Neural Networks for Semantic Segmentation
Found it because it cite "Deep neural networks segment neuronal membranes in electron microscopy images", and it also cite the "U-Net" paper.
Exact mapping from random forest to CNN ,which can be used to initialize CNN with sparse training data. Approximate mapping from CNN to RF, and the RF improve original RF.
Shi Jianbo's second author. Also check out recent Arxiv paper for edge detection.
Reformulate mean field approximation by recurrent network. that seems a reminder that old problem (CRF, graph-cut etc) can be re-formulated into a neural network optimization problem, or be re-formulated such that it can be optimized by gradient together with the deep network. Need to read this work.
The dense CRF paper. Define unary potential is defined from conventional features like color, texture or spatial location, and the pairwise potentials are define similar to Boykov: a indicator function for the equality of labels, and a mixture of Gaussian kernels as function of features. Here the neighborhood is not local any more, but includes the whole image field. Naive variational inference need update each pixel with cost of O(N). This paper used an approximation: down sampling image, convolution with kernels, and then upsampling. Also used another approximation called permutohedral lattice so the computation is linear with dimension d.
Code is available on author's website with python wrapper (which seems have some functions for > 2D images. )
This paper address the problem of segmentation given only scribble about the interior of regions. The problem is similar to Boykov's graph cut, and the optimization also include graph-cut. The unary potential includes two parts: one is from the scribble, and another is from the trained CNN. The pairwise potential is defined like Boykov. The graph is defined on superpixels. The optimization alternates between two steps: 1) Given CNN fixed, optimize graph cut problem on superpixels. 2) Given the segmentation labels of the superpixels estimated from graph-cut, train a fully CNN on each pixel.
One of the major CNN+CRF papers.
Textonboost for image understanding: Multi-class object recognition and segmentation by jointly modeling texture
Seems a major work before deep learning, for feature engineering and detection.
Currently the PASCSL VOC leaderboard top. Use potentials that include > 2 nodes. 1) use a object detector and use the detection results (foreground after grabcut) as potentials and 2) use superpixel as higher order potential.
Used mean field update, with fully connected CRF. But the GPU algorithm is not clear to me.