ICCV 2015 Paper Navigator

maintained by: colin lea
original concept by: andrej karpathy
Other guides:
CVPR: 2015, 2014, 2013 ICCV: 2013
NIPS: 2015, 2014, 2013, 2012 ICML: 2013
All 526 ICCV papers are clustered into the topics below using TF-IDF similarity on the words in each paper.
Each color indicates a different cluster index.
Toggle the topics below to sort by category. The top 10 words from each cluster are shown.

Filter papers (e.g. by keyword or author):
Discrete Tabu Search for Graph Matching [pdf]
Kamil Adamczewski, Yumin Suh, Kyoung Mu Lee

Abstract: Graph matching is a fundamental problem in computer vision. In this paper, we propose a novel graph matching algorithm based on tabu search [13]. The proposed method solves graph matching problem by casting it into an equivalent weighted maximum clique problem of the corresponding association graph, which we further penalize through introducing negative weights. Subsequent tabu search optimization allows for overcoming the convention of using positive weights. The methods distinct feature is that it utilizes the history of search to make more strategic decisions while looking for the optimal solution, thus effectively escaping local optima and in practice achieving superior results. The proposed method, unlike the existing algorithms, enables direct optimization in the original discrete space while encouraging rather than artificially enforcing hard one-to-one constraint, thus resulting in better solution. The experiments demonstrate the robustness of the algorithm in a variety of settings, presenting the state-of-the-art results. The code is available at http://cv.snu.ac.kr/research/DTSGM/.
Similar papers:
  • Efficient Decomposition of Image and Mesh Graphs by Lifted Multicuts [pdf] - Margret Keuper, Evgeny Levinkov, Nicolas Bonneel, Guillaume Lavoue, Thomas Brox, Bjorn Andres
  • Multi-Image Matching via Fast Alternating Minimization [pdf] - Xiaowei Zhou, Menglong Zhu, Kostas Daniilidis
  • Attribute-Graph: A Graph Based Approach to Image Ranking [pdf] - Nikita Prabhu, R. Venkatesh Babu
  • Mining And-Or Graphs for Graph Matching and Object Discovery [pdf] - Quanshi Zhang, Ying Nian Wu, Song-Chun Zhu
  • A Matrix Decomposition Perspective to Multiple Graph Matching [pdf] - Junchi Yan, Hongteng Xu, Hongyuan Zha, Xiaokang Yang, Huanxi Liu, Stephen Chu
Learning to See by Moving [pdf]
Pulkit Agrawal, Joao Carreira, Jitendra Malik

Abstract: The current dominant paradigm for feature learning in computer vision relies on training neural networks for the task of object recognition using millions of hand labelled images. Is it also possible to learn features for a diverse set of visual tasks using any other form of supervision? In biology, living organisms developed the ability of visual perception for the purpose of moving and acting in the world. Drawing inspiration from this observation, in this work we investigated if the awareness of egomotion (i.e. self motion) can be used as a supervisory signal for feature learning. As opposed to the knowledge of class labels, information about egomotion is freely available to mobile agents. We found that using the same number of training images, features learnt using egomotion as supervision compare favourably to features learnt using class-label as supervision on the tasks of scene recognition, object recognition, visual odometry and keypoint matching. We must perceive in order to move, but we must also move in order to perceive J.J Gibson
Similar papers:
  • Learning The Structure of Deep Convolutional Networks [pdf] - Jiashi Feng, Trevor Darrell
  • Domain Generalization for Object Recognition With Multi-Task Autoencoders [pdf] - Muhammad Ghifary, W. Bastiaan Kleijn, Mengjie Zhang, David Balduzzi
  • Unsupervised Visual Representation Learning by Context Prediction [pdf] - Carl Doersch, Abhinav Gupta, Alexei A. Efros
  • Unsupervised Learning of Visual Representations Using Videos [pdf] - Xiaolong Wang, Abhinav Gupta
  • Learning Image Representations Tied to Ego-Motion [pdf] - Dinesh Jayaraman, Kristen Grauman
Learning Shape, Motion and Elastic Models in Force Space [pdf]
Antonio Agudo, Francesc Moreno-Noguer

Abstract: In this paper, we address the problem of simultaneously recovering the 3D shape and pose of a deformable and potentially elastic object from 2D motion. This is a highly ambiguous problem typically tackled by using low-rank shape and trajectory constraints. We show that formulating the problem in terms of a low-rank force space that induces the deformation, allows for a better physical interpretation of the resulting priors and a more accurate representation of the actual objects behavior. However, this comes at the price of, besides force and pose, having to estimate the elastic model of the object. For this, we use an Expectation Maximization strategy, where each of these parameters are successively learned within partial M-steps, while robustly dealing with missing observations. We thoroughly validate the approach on both mocap and real sequences, showing more accurate 3D reconstructions than state-of-the-art, and additionally providing an estimate of the full elastic model with no a priori information.
Similar papers:
  • Sparse Dynamic 3D Reconstruction From Unsynchronized Videos [pdf] - Enliang Zheng, Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm
  • Detailed Full-Body Reconstructions of Moving People From Monocular RGB-D Sequences [pdf] - Federica Bogo, Michael J. Black, Matthew Loper, Javier Romero
  • The One Triangle Three Parallelograms Sampling Strategy and Its Application in Shape Regression [pdf] - Mikael Nilsson
  • Shell PCA: Statistical Shape Modelling in Shell Space [pdf] - Chao Zhang, Behrend Heeren, Martin Rumpf, William A. P. Smith
  • Compositional Hierarchical Representation of Shape Manifolds for Classification of Non-Manifold Shapes [pdf] - Mete Ozay, Umit Rusen Aktas, Jeremy L. Wyatt, Ales Leonardis
Optimizing Expected Intersection-Over-Union With Candidate-Constrained CRFs [pdf]
Faruk Ahmed, Dany Tarlow, Dhruv Batra

Abstract: We study the question of how to make loss-aware predictions in image segmentation settings where the evaluation function is the Intersection-over-Union (IoU) measure that is used widely in evaluating image segmentation systems. Currently, there are two dominant approaches: the first approximates the Expected-IoU (EIoU) score as ExpectedIntersection-over-Expected-Union (EIoEU); and the second approach is to compute exact EIoU but only over a small set of high-quality candidate solutions. We begin by asking which approach we should favor for two typical image segmentation tasks. Studying this question leads to two new methods that draw ideas from both existing approaches. Our new methods use the EIoEU approximation paired with high quality candidate solutions. Experimentally we show that our new approaches lead to improved performance on both image segmentation tasks.
Similar papers:
  • Predicting Multiple Structured Visual Interpretations [pdf] - Debadeepta Dey, Varun Ramakrishna, Martial Hebert, J. Andrew Bagnell
  • An NMF Perspective on Binary Hashing [pdf] - Lopamudra Mukherjee, Sathya N. Ravi, Vamsi K. Ithapu, Tyler Holmes, Vikas Singh
  • Discrete Tabu Search for Graph Matching [pdf] - Kamil Adamczewski, Yumin Suh, Kyoung Mu Lee
  • A Randomized Ensemble Approach to Industrial CT Segmentation [pdf] - Hyojin Kim, Jayaraman Jayaraman J. Thiagarajan, Peer-Timo Bremer
  • Inferring M-Best Diverse Labelings in a Single One [pdf] - Alexander Kirillov, Bogdan Savchynskyy, Dmitrij Schlesinger, Dmitry Vetrov, Carsten Rother
RGB-W: When Vision Meets Wireless [pdf]
Alexandre Alahi, Albert Haque, Li Fei-Fei

Abstract: Meters 20 Inspired by the recent success of RGB-D cameras, we propose the enrichment of RGB data with an additional quasi-free modality, namely, the wireless signal emitted by individuals cell phones, referred to as RGB-W. The received signal strength acts as a rough proxy for depth and a reliable cue on a persons identity. Although the measured signals are noisy, we demonstrate that the combination of visual and wireless data significantly improves the localization accuracy. We introduce a novel image-driven representation of wireless data which embeds all received signals onto a single image. We then evaluate the ability of this additional data to (i) locate persons within a sparsitydriven framework and to (ii) track individuals with a new confidence measure on the data association problem. Our solution outperforms existing localization methods. It can be applied to the millions of currently installed RGB cameras to better analyze human behavior and offer the next generation of high-accuracy location-based services.
Similar papers:
  • Learning to Track for Spatio-Temporal Action Localization [pdf] - Philippe Weinzaepfel, Zaid Harchaoui, Cordelia Schmid
  • A Versatile Learning-Based 3D Temporal Tracker: Scalable, Robust, Online [pdf] - David Joseph Tan, Federico Tombari, Slobodan Ilic, Nassir Navab
  • Joint Probabilistic Data Association Revisited [pdf] - Seyed Hamid Rezatofighi, Anton Milan, Zhen Zhang, Qinfeng Shi, Anthony Dick, Ian Reid
  • Lost Shopping! Monocular Localization in Large Indoor Spaces [pdf] - Shenlong Wang, Sanja Fidler, Raquel Urtasun
  • You Are Here: Mimicking the Human Thinking Process in Reading Floor-Plans [pdf] - Hang Chu, Dong Ki Kim, Tsuhan Chen
Learning to Transfer: Transferring Latent Task Structures and Its Application to Person-Specific Facial Action Unit Detection [pdf]
Timur Almaev, Brais Martinez, Michel Valstar

Abstract: In this article we explore the problem of constructing person-specific models for the detection of facial Action Units (AUs), addressing the problem from the point of view of Transfer Learning and Multi-Task Learning. Our starting point is the fact that some expressions, such as smiles, are very easily elicited, annotated, and automatically detected, while others are much harder to elicit and to annotate. We thus consider a novel problem: all AU models for the target subject are to be learnt using person-specific annotated data for a reference AU (AU12 in our case), and no data or little data regarding the target AU. In order to design such a model, we propose a novel Multi-Task Learning and the associated Transfer Learning framework, in which we consider both relations across subjects and AUs. That is to say, we consider a tensor structure among the tasks. Our approach hinges on learning the latent relations among tasks using one single reference AU, and then transferring these latent relations to other AUs. We show that we are able to effectively make use of the annotated data for AU12 when learning other person-specific AU models, even in the absence of data for the target task. Finally, we show the excellent performance of our method when small amounts of annotated data for the target tasks are made available.
Similar papers:
  • Pairwise Conditional Random Forests for Facial Expression Recognition [pdf] - Arnaud Dapogny, Kevin Bailly, Severine Dubuisson
  • FaceDirector: Continuous Control of Facial Performance in Video [pdf] - Charles Malleson, Jean-Charles Bazin, Oliver Wang, Derek Bradley, Thabo Beeler, Adrian Hilton, Alexander Sorkine-Hornung
  • Joint Fine-Tuning in Deep Neural Networks for Facial Expression Recognition [pdf] - Heechul Jung, Sihaeng Lee, Junho Yim, Sunjeong Park, Junmo Kim
  • Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf] - Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic
  • From Emotions to Action Units With Hidden and Semi-Hidden-Task Learning [pdf] - Adria Ruiz, Joost Van de Weijer, Xavier Binefa
Contractive Rectifier Networks for Nonlinear Maximum Margin Classification [pdf]
Senjian An, Munawar Hayat, Salman H. Khan, Mohammed Bennamoun, Farid Boussaid, Ferdous Sohel

Abstract: To find the optimal nonlinear separating boundary with maximum margin in the input data space, this paper proposes Contractive Rectifier Networks (CRNs), wherein the hidden-layer transformations are restricted to be contraction mappings. The contractive constraints ensure that the achieved separating margin in the input space is larger than or equal to the separating margin in the output layer. The training of the proposed CRNs is formulated as a linear support vector machine (SVM) in the output layer, combined with two or more contractive hidden layers. Effective algorithms have been proposed to address the optimization challenges arising from contraction constraints. Experimental results on MNIST, CIFAR-10, CIFAR-100 and MIT67 datasets demonstrate that the proposed contractive rectifier networks consistently outperform their conventional unconstrained rectifier network counterparts.
Similar papers:
  • Multi-Scale Recognition With DAG-CNNs [pdf] - Songfan Yang, Deva Ramanan
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • Learning The Structure of Deep Convolutional Networks [pdf] - Jiashi Feng, Trevor Darrell
  • An Exploration of Parameter Redundancy in Deep Networks With Circulant Projections [pdf] - Yu Cheng, Felix X. Yu, Rogerio S. Feris, Sanjiv Kumar, Alok Choudhary, Shi-Fu Chang
  • Scalable Nonlinear Embeddings for Semantic Category-Based Image Retrieval [pdf] - Gaurav Sharma, Bernt Schiele
Per-Sample Kernel Adaptation for Visual Recognition and Grouping [pdf]
Borislav Antic, Bjorn Ommer

Abstract: Object, action, or scene representations that are corrupted by noise significantly impair the performance of visual recognition. Typically, partial occlusion, clutter, or excessive articulation affects only a subset of all feature dimensions and, most importantly, different dimensions are corrupted in different samples. Nevertheless, the common approach to this problem in feature selection and kernel methods is to down-weight or eliminate entire training samples or the same dimensions of all samples. Thus, valuable signal is lost, resulting in suboptimal classification. Our goal is, therefore, to adjust the contribution of individual feature dimensions when comparing any two samples and computing their similarity. Consequently, persample selection of informative dimensions is directly integrated into kernel computation. The interrelated problems of learning the parameters of a kernel classifier and determining the informative components of each sample are then addressed in a joint objective function. The approach can be integrated into the learning stage of any kernel-based visual recognition problem and it does not affect the computational performance in the retrieval phase. Experiments on diverse challenges of action recognition in videos and indoor scene classification show the general applicability of the approach and its ability to improve learning of visual representations.
Similar papers:
  • Human Action Recognition Using Factorized Spatio-Temporal Convolutional Networks [pdf] - Lin Sun, Kui Jia, Dit-Yan Yeung, Bertram E. Shi
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Low Dimensional Explicit Feature Maps [pdf] - Ondrej Chum
  • Bayesian Model Adaptation for Crowd Counts [pdf] - Bo Liu, Nuno Vasconcelos
  • Beyond Covariance: Feature Representation With Nonlinear Kernel Matrices [pdf] - Lei Wang, Jianjia Zhang, Luping Zhou, Chang Tang, Wanqing Li
VQA: Visual Question Answering [pdf]
Stanislaw Antol, Aishwarya Agrawal, Jiasen Lu, Margaret Mitchell, Dhruv Batra, C. Lawrence Zitnick, Devi Parikh

Abstract: We propose the task of free-form and open-ended Visual Question Answering (VQA). Given an image and a natural language question about the image, the task is to provide an accurate natural language answer. Mirroring real-world scenarios, such as helping the visually impaired, both the questions and answers are open-ended. Visual questions selectively target different areas of an image, including background details and underlying context. As a result, a system that succeeds at VQA typically needs a more detailed understanding of the image and complex reasoning than a system producing generic image captions. Moreover, VQA is amenable to automatic evaluation, since many open-ended answers contain only a few words or a closed set of answers that can be provided in a multiple-choice format. We provide a dataset containing 0.25M images, 0.76M questions, and 10M answers (www.visualqa.org), and discuss the information it provides. Numerous baselines for VQA are provided and compared with human performance.
Similar papers:
  • Common Subspace for Model and Similarity: Phrase Learning for Caption Generation From Images [pdf] - Yoshitaka Ushiku, Masataka Yamaguchi, Yusuke Mukuta, Tatsuya Harada
  • A Data-Driven Metric for Comprehensive Evaluation of Saliency Models [pdf] - Jia Li, Changqun Xia, Yafei Song, Shu Fang, Xiaowu Chen
  • Learning Common Sense Through Visual Abstraction [pdf] - Ramakrishna Vedantam, Xiao Lin, Tanmay Batra, C. Lawrence Zitnick, Devi Parikh
  • Visual Madlibs: Fill in the Blank Description Generation and Question Answering [pdf] - Licheng Yu, Eunbyung Park, Alexander C. Berg, Tamara L. Berg
  • Ask Your Neurons: A Neural-Based Approach to Answering Questions About Images [pdf] - Mateusz Malinowski, Marcus Rohrbach, Mario Fritz
Class-Specific Image Deblurring [pdf]
Saeed Anwar, Cong Phuoc Huynh, Fatih Porikli

Abstract: In image deblurring, a fundamental problem is that the blur kernel suppresses a number of spatial frequencies that are difficult to recover reliably. In this paper, we explore the potential of a class-specific image prior for recovering spatial frequencies attenuated by the blurring process. Specifically, we devise a prior based on the class-specific subspace of image intensity responses to band-pass filters. We learn that the aggregation of these subspaces across all frequency bands serves as a good class-specific prior for the restoration of frequencies that cannot be recovered with generic image priors. In an extensive validation, our method, equipped with the above prior, yields greater image quality than many state-of-the-art methods by up to 5 dB in terms of image PSNR, across various image categories including portraits, cars, cats, pedestrians and household objects.
Similar papers:
  • Per-Sample Kernel Adaptation for Visual Recognition and Grouping [pdf] - Borislav Antic, Bjorn Ommer
  • Self-Calibration of Optical Lenses [pdf] - Michael Hirsch, Bernhard Scholkopf
  • Low Dimensional Explicit Feature Maps [pdf] - Ondrej Chum
  • Conditioned Regression Models for Non-Blind Single Image Super-Resolution [pdf] - Gernot Riegler, Samuel Schulter, Matthias Ruther, Horst Bischof
  • Complementary Sets of Shutter Sequences for Motion Deblurring [pdf] - Hae-Gon Jeon, Joon-Young Lee, Yudeog Han, Seon Joo Kim, In So Kweon
Understanding Deep Features With Computer-Generated Imagery [pdf]
Mathieu Aubry, Bryan C. Russell

Abstract: We introduce an approach for analyzing the variation of features generated by convolutional neural networks (CNNs) with respect to scene factors that occur in natural images. Such factors may include object style, 3D viewpoint, color, and scene lighting configuration. Our approach analyzes CNN feature responses corresponding to different scene factors by controlling for them via rendering using a large database of 3D CAD models. The rendered images are presented to a trained CNN and responses for different layers are studied with respect to the input scene factors. We perform a decomposition of the responses based on knowledge of the input scene factors and analyze the resulting components. In particular, we quantify their relative importance in the CNN responses and visualize them using principal component analysis. We show qualitative and quantitative results of our study on three CNNs trained on large image datasets: AlexNet [18], Places [43], and Oxford VGG [8]. We observe important differences across the networks and CNN layers for different scene factors and object categories. Finally, we demonstrate that our analysis based on computergenerated imagery translates to the network representation of natural images.
Similar papers:
  • Matrix Backpropagation for Deep Networks With Structured Layers [pdf] - Catalin Ionescu, Orestis Vantzos, Cristian Sminchisescu
  • Unsupervised Generation of a Viewpoint Annotated Car Dataset From Videos [pdf] - Nima Sedaghat, Thomas Brox
  • Single Image 3D Without a Single 3D Image [pdf] - David F. Fouhey, Wajahat Hussain, Abhinav Gupta, Martial Hebert
  • Learning Visual Clothing Style With Heterogeneous Dyadic Co-Occurrences [pdf] - Andreas Veit, Balazs Kovacs, Sean Bell, Julian McAuley, Kavita Bala, Serge Belongie
  • Render for CNN: Viewpoint Estimation in Images Using CNNs Trained With Rendered 3D Model Views [pdf] - Hao Su, Charles R. Qi, Yangyan Li, Leonidas J. Guibas
Web-Scale Image Clustering Revisited [pdf]
Yannis Avrithis, Yannis Kalantidis, Evangelos Anagnostopoulos, Ioannis Z. Emiris

Abstract: Large scale duplicate detection, clustering and mining of documents or images has been conventionally treated with seed detection via hashing, followed by seed growing heuristics using fast search. Principled clustering methods, especially kernelized and spectral ones, have higher complexity and are difficult to scale above millions. Under the assumption of documents or images embedded in Euclidean space, we revisit recent advances in approximate k-means variants, and borrow their best ingredients to introduce a new one, inverted-quantized k-means (IQ-means). Key underlying concepts are quantization of data points and multi-index based inverted search from centroids to cells. Its quantization is a form of hashing and analogous to seed detection, while its updates are analogous to seed growing, yet principled in the sense of distortion minimization. We further design a dynamic variant that is able to determine the number of clusters k in a single run at nearly zero additional cost. Combined with powerful deep learned representations, we achieve clustering of a 100 million image collection on a single machine in less than one hour. (a) Ranked retrieval [8] (b) DRVQ [1] (c) EGM [2] (d) This work: IQ-means Figure 1. Different k-means variants. ( ) Data points; ( ) centroids; ( ) search range; ( ) estimated cluster extent, used to dynamically determine k.
Similar papers:
  • Scalable Person Re-Identification: A Benchmark [pdf] - Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian
  • Low-Rank Tensor Constrained Multiview Subspace Clustering [pdf] - Changqing Zhang, Huazhu Fu, Si Liu, Guangcan Liu, Xiaochun Cao
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
  • Multi-View Subspace Clustering [pdf] - Hongchang Gao, Feiping Nie, Xuelong Li, Heng Huang
  • Temporal Subspace Clustering for Human Motion Segmentation [pdf] - Sheng Li, Kang Li, Yun Fu
Predicting Deep Zero-Shot Convolutional Neural Networks Using Textual Descriptions [pdf]
Jimmy Lei Ba, Kevin Swersky, Sanja Fidler, Ruslan salakhutdinov

Abstract: One of the main challenges in Zero-Shot Learning of visual categories is gathering semantic attributes to accompany images. Recent work has shown that learning from textual descriptions, such as Wikipedia articles, avoids the problem of having to explicitly define these attributes. We present a new model that can classify unseen categories from their textual description. Specifically, we use text features to predict the output weights of both the convolutional and the fully connected layers in a deep convolutional neural network (CNN). We take advantage of the architecture of CNNs and learn features at different layers, rather than just learning an embedding space for both modalities, as is common with existing approaches. The proposed model also allows us to automatically generate a list of pseudoattributes for each visual category consisting of words from Wikipedia articles. We train our models end-to-end using the Caltech-UCSD bird and flower datasets and evaluate both ROC and Precision-Recall curves. Our empirical results show that the proposed model significantly outperforms previous methods.
Similar papers:
  • FASText: Efficient Unconstrained Scene Text Detector [pdf] - Michal Busta, Lukas Neumann, Jiri Matas
  • Unsupervised Domain Adaptation for Zero-Shot Learning [pdf] - Elyor Kodirov, Tao Xiang, Zhenyong Fu, Shaogang Gong
  • Text Flow: A Unified Text Detection System in Natural Scene Images [pdf] - Shangxuan Tian, Yifeng Pan, Chang Huang, Shijian Lu, Kai Yu, Chew Lim Tan
  • Semi-Supervised Zero-Shot Classification With Label Representation Learning [pdf] - Xin Li, Yuhong Guo, Dale Schuurmans
  • Zero-Shot Learning via Semantic Similarity Embedding [pdf] - Ziming Zhang, Venkatesh Saligrama
Aggregating Local Deep Features for Image Retrieval [pdf]
Artem Babenko, Victor Lempitsky

Abstract: Several recent works have shown that image descriptors produced by deep convolutional neural networks provide state-of-the-art performance for image classification and retrieval problems. It has also been shown that the activations from the convolutional layers can be interpreted as local features describing particular image regions. These local features can be aggregated using aggregation approaches developed for local features (e.g. Fisher vectors), thus providing new powerful global descriptors. In this paper we investigate possible ways to aggregate local deep features to produce compact global descriptors for image retrieval. First, we show that deep features and traditional hand-engineered features have quite different distributions of pairwise similarities, hence existing aggregation methods have to be carefully re-evaluated. Such re-evaluation reveals that in contrast to shallow features, the simple aggregation method based on sum pooling provides arguably the best performance for deep convolutional features. This method is efficient, has few parameters, and bears little risk of overfitting when e.g. learning the PCA matrix. Overall, the new compact global descriptor improves the state-of-the-art on four common benchmarks considerably.
Similar papers:
  • Deep Multi-Patch Aggregation Network for Image Style, Aesthetics, and Quality Estimation [pdf] - Xin Lu, Zhe Lin, Xiaohui Shen, Radomir Mech, James Z. Wang
  • Discriminative Learning of Deep Convolutional Feature Point Descriptors [pdf] - Edgar Simo-Serra, Eduard Trulls, Luis Ferraz, Iasonas Kokkinos, Pascal Fua, Francesc Moreno-Noguer
  • Multi-View Convolutional Neural Networks for 3D Shape Recognition [pdf] - Hang Su, Subhransu Maji, Evangelos Kalogerakis, Erik Learned-Miller
  • Local Convolutional Features With Unsupervised Training for Image Retrieval [pdf] - Mattis Paulin, Matthijs Douze, Zaid Harchaoui, Julien Mairal, Florent Perronin, Cordelia Schmid
  • RIDE: Reversal Invariant Descriptor Enhancement [pdf] - Lingxi Xie, Jingdong Wang, Weiyao Lin, Bo Zhang, Qi Tian
Flow Fields: Dense Correspondence Fields for Highly Accurate Large Displacement Optical Flow Estimation [pdf]
Christian Bailer, Bertram Taetz, Didier Stricker

Abstract: Modern large displacement optical flow algorithms usually use an initialization by either sparse descriptor matching techniques or dense approximate nearest neighbor fields. While the latter have the advantage of being dense, they have the major disadvantage of being very outlier prone as they are not designed to find the optical flow, but the visually most similar correspondence. In this paper we present a dense correspondence field approach that is much less outlier prone and thus much better suited for optical flow estimation than approximate nearest neighbor fields. Our approach is conceptually novel as it does not require explicit regularization, smoothing (like median filtering) or a new data term, but solely our novel purely data based search strategy that finds most inliers (even for small objects), while it effectively avoids finding outliers. Moreover, we present novel enhancements for outlier filtering. We show that our approach is better suited for large displacement optical flow estimation than state-of-the-art descriptor matching techniques. We do so by initializing EpicFlow (so far the best method on MPI-Sintel) with our Flow Fields instead of their originally used state-of-the-art descriptor matching technique. We significantly outperform the original EpicFlow on MPI-Sintel, KITTI and Middlebury.
Similar papers:
  • SPM-BP: Sped-up PatchMatch Belief Propagation for Continuous MRFs [pdf] - Yu Li, Dongbo Min, Michael S. Brown, Minh N. Do, Jiangbo Lu
  • Large Displacement 3D Scene Flow With Occlusion Reasoning [pdf] - Andrei Zanfir, Cristian Sminchisescu
  • Face Flow [pdf] - Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou
  • FlowNet: Learning Optical Flow With Convolutional Networks [pdf] - Alexey Dosovitskiy, Philipp Fischer, Eddy Ilg, Philip Hausser, Caner Hazirbas, Vladimir Golkov, Patrick van der Smagt, Daniel Cremers, Thomas Brox
  • Oriented Light-Field Windows for Scene Flow [pdf] - Pratul P. Srinivasan, Michael W. Tao, Ren Ng, Ravi Ramamoorthi
Lending A Hand: Detecting Hands and Recognizing Activities in Complex Egocentric Interactions [pdf]
Sven Bambach, Stefan Lee, David J. Crandall, Chen Yu

Abstract: Hands appear very often in egocentric video, and their appearance and pose give important cues about what people are doing and what they are paying attention to. But existing work in hand detection has made strong assumptions that work well in only simple scenarios, such as with limited interaction with other people or in lab settings. We develop methods to locate and distinguish between hands in egocentric video using strong appearance models with Convolutional Neural Networks, and introduce a simple candidate region generation approach that outperforms existing techniques at a fraction of the computational cost. We show how these high-quality bounding boxes can be used to create accurate pixelwise hand regions, and as an application, we investigate the extent to which hand segmentation alone can distinguish between different activities. We evaluate these techniques on a new dataset of 48 first-person videos of people interacting in realistic environments, with pixel-level ground truth for over 15,000 hand instances.
Similar papers:
  • 3D Hand Pose Estimation Using Randomized Decision Forest With Segmentation Index Points [pdf] - Peiyi Li, Haibin Ling, Xi Li, Chunyuan Liao
  • A Collaborative Filtering Approach to Real-Time Hand Pose Estimation [pdf] - Chiho Choi, Ayan Sinha, Joon Hee Choi, Sujin Jang, Karthik Ramani
  • 3D Object Reconstruction From Hand-Object Interactions [pdf] - Dimitrios Tzionas, Juergen Gall
  • Depth-Based Hand Pose Estimation: Data, Methods, and Challenges [pdf] - James S. Supancic III, Gregory Rogez, Yi Yang, Jamie Shotton, Deva Ramanan
  • Understanding Everyday Hands in Action From RGB-D Images [pdf] - Gregory Rogez, James S. Supancic III, Deva Ramanan
Contour Detection and Characterization for Asynchronous Event Sensors [pdf]
Francisco Barranco, Ching L. Teo, Cornelia Fermuller, Yiannis Aloimonos

Abstract: The bio-inspired, asynchronous event-based dynamic vision sensor records temporal changes in the luminance of the scene at high temporal resolution. Since events are only triggered at significant luminance changes, most events occur at the boundary of objects and their parts. The detection of these contours is an essential step for further interpretation of the scene. This paper presents an approach to learn the location of contours and their border ownership using Structured Random Forests on event-based features that encode motion, timing, texture, and spatial orientations. The classifier integrates elegantly information over time by utilizing the classification results previously computed. Finally, the contour detection and boundary assignment are demonstrated in a layer-segmentation of the scene. Experimental results demonstrate good performance in boundary detection and segmentation.
Similar papers:
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • Multiresolution Hierarchy Co-Clustering for Semantic Segmentation in Sequences With Small Variations [pdf] - David Varas, Monica Alfaro, Ferran Marques
  • Motion Trajectory Segmentation via Minimum Cost Multicuts [pdf] - Margret Keuper, Bjoern Andres, Thomas Brox
  • High-for-Low and Low-for-High: Efficient Boundary Detection From Deep Object Features and its Applications to High-Level Vision [pdf] - Gedas Bertasius, Jianbo Shi, Lorenzo Torresani
  • Contour Flow: Middle-Level Motion Estimation by Combining Motion Segmentation and Contour Alignment [pdf] - Huijun Di, Qingxuan Shi, Feng Lv, Ming Qin, Yao Lu
Convolutional Color Constancy [pdf]
Jonathan T. Barron

Abstract: Color constancy is the problem of inferring the color of the light that illuminated a scene, usually so that the illumination color can be removed. Because this problem is underconstrained, it is often solved by modeling the statistical regularities of the colors of natural objects and illumination. In contrast, in this paper we reformulate the problem of color constancy as a 2D spatial localization task in a logchrominance space, thereby allowing us to apply techniques from object detection and structured prediction to the color constancy problem. By directly learning how to discriminate between correctly white-balanced images and poorly white-balanced images, our model is able to improve performance on standard benchmarks by nearly 40%. assumes that the illuminant color is the average color of all image pixels, thereby implicitly assuming that object reflectances are, on average, gray [12]. This simple idea can be generalized to modeling gradient information or using generalized norms instead of a simple arithmetic mean [4, 36], modeling the spatial distribution of colors with a filter bank [13], modeling the distribution of color histograms [21], or implicitly reasoning about the moments of colors using PCA [14]. Other models assume that the colors of natural objects lie with some gamut [3, 23]. Most of these models can be thought of as statistical, as they either assume some distribution of colors or they learn some distribution of colors from training data. This connect
Similar papers:
  • Illumination Robust Color Naming via Label Propagation [pdf] - Yuanliu liu, Zejian Yuan, Badong Chen, Jianru Xue, Nanning Zheng
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
  • Learning Large-Scale Automatic Image Colorization [pdf] - Aditya Deshpande, Jason Rock, David Forsyth
  • Beyond White: Ground Truth Colors for Color Constancy Correction [pdf] - Dongliang Cheng, Brian Price, Scott Cohen, Michael S. Brown
A Comprehensive Multi-Illuminant Dataset for Benchmarking of the Intrinsic Image Algorithms [pdf]
Shida Beigpour, Andreas Kolb, Sven Kunz

Abstract: In this paper, we provide a new, real photo dataset with precise ground-truth for intrinsic image research. Prior ground-truth datasets have been restricted to rather simple illumination conditions and scene geometries, or have been enhanced using image synthesis methods. The dataset provided in this paper is based on complex multi-illuminant scenarios under multi-colored illumination conditions and challenging cast shadows. We provide full per-pixel intrinsic ground-truth data for these scenarios, i.e. reflectance, specularity, shading, and illumination for scenes as well as preliminary depth information. Furthermore, we evaluate 3 state-of-the-art intrinsic image recovery methods, using our dataset.
Similar papers:
  • Direct Intrinsics: Learning Albedo-Shading Decomposition by Convolutional Regression [pdf] - Takuya Narihira, Michael Maire, Stella X. Yu
  • Learning Ordinal Relationships for Mid-Level Vision [pdf] - Daniel Zoran, Phillip Isola, Dilip Krishnan, William T. Freeman
  • Learning Data-Driven Reflectance Priors for Intrinsic Image Decomposition [pdf] - Tinghui Zhou, Philipp Krahenbuhl, Alexei A. Efros
  • Intrinsic Decomposition of Image Sequences From Local Temporal Variations [pdf] - Pierre-Yves Laffont, Jean-Charles Bazin
  • Intrinsic Scene Decomposition From RGB-D images [pdf] - Mohammed Hachama, Bernard Ghanem, Peter Wonka
Robust Optimization for Deep Regression [pdf]
Vasileios Belagiannis, Christian Rupprecht, Gustavo Carneiro, Nassir Navab

Abstract: L2 loss Tukey's biweight loss 70 50 40 30 20 10 0 PARSE LSP Football Volleyball Dataset Error 8 L2 loss Tukey's biweight loss 7 Mean Pixel Error (MPE) Convolutional Neural Networks (ConvNets) have successfully contributed to improve the accuracy of regressionbased methods for computer vision tasks such as human pose estimation, landmark localization, and object detection. The network optimization has been usually performed with L2 loss and without considering the impact of outliers on the training process, where an outlier in this context is defined by a sample estimation that lies at an abnormal distance from the other training sample estimations in the objective space. In this work, we propose a regression model with ConvNets that achieves robustness to such outliers by minimizing Tukeys biweight function, an M-estimator robust to outliers, as the loss function for the ConvNet. In addition to the robust loss, we introduce a coarse-to-fine model, which processes input images of progressively higher resolutions for improving the accuracy of the regressed values. In our experiments, we demonstrate faster convergence and better generalization of our robust loss function for the tasks of human pose estimation and age estimation from face images. We also show that the combination of the robust loss function with the coarse-to-fine model produces comparable or better results than current state-of-the-art approaches in four publicly available human pose estimation dat
Similar papers:
  • Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf] - Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • Uncovering Interactions and Interactors: Joint Estimation of Head, Body Orientation and F-Formations From Surveillance Videos [pdf] - Elisa Ricci, Jagannadan Varadarajan, Ramanathan Subramanian, Samuel Rota Bulo, Narendra Ahuja, Oswald Lanz
  • Flowing ConvNets for Human Pose Estimation in Videos [pdf] - Tomas Pfister, James Charles, Andrew Zisserman
  • PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization [pdf] - Alex Kendall, Matthew Grimes, Roberto Cipolla
High-for-Low and Low-for-High: Efficient Boundary Detection From Deep Object Features and its Applications to High-Level Vision [pdf]
Gedas Bertasius, Jianbo Shi, Lorenzo Torresani

Abstract: Most of the current boundary detection systems rely exclusively on low-level features, such as color and texture. However, perception studies suggest that humans employ object-level reasoning when judging if a particular pixel is a boundary. Inspired by this observation, in this work we show how to predict boundaries by exploiting objectlevel features from a pretrained object-classification network. Our method can be viewed as a High-for-Low approach where high-level object features inform the low-level boundary detection process. Our model achieves state-ofthe-art performance on an established boundary detection benchmark and it is efficient to run. Additionally, we show that due to the semantic nature of our boundaries we can use them to aid a number of highlevel vision tasks. We demonstrate that by using our boundaries we improve the performance of state-of-the-art methods on the problems of semantic boundary labeling, semantic segmentation and object proposal generation. We can view this process as a Low-for-High scheme, where lowlevel boundaries aid high-level vision tasks. Thus, our contributions include a boundary detection system that is accurate, efficient, generalizes well to multiple datasets, and is also shown to improve existing stateof-the-art high-level vision methods on three distinct tasks.
Similar papers:
  • Efficient Classifier Training to Minimize False Merges in Electron Microscopy Segmentation [pdf] - Toufiq Parag, Dan C. Ciresan, Alessandro Giusti
  • Projection Onto the Manifold of Elongated Structures for Accurate Extraction [pdf] - Amos Sironi, Vincent Lepetit, Pascal Fua
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
  • Contour Detection and Characterization for Asynchronous Event Sensors [pdf] - Francisco Barranco, Ching L. Teo, Cornelia Fermuller, Yiannis Aloimonos
  • Multiresolution Hierarchy Co-Clustering for Semantic Segmentation in Sequences With Small Variations [pdf] - David Varas, Monica Alfaro, Ferran Marques
Detailed Full-Body Reconstructions of Moving People From Monocular RGB-D Sequences [pdf]
Federica Bogo, Michael J. Black, Matthew Loper, Javier Romero

Abstract: We accurately estimate the 3D geometry and appearance of the human body from a monocular RGB-D sequence of a user moving freely in front of the sensor. Range data in each frame is first brought into alignment with a multi-resolution 3D body model in a coarse-to-fine process. The method then uses geometry and image texture over time to obtain accurate shape, pose, and appearance information despite unconstrained motion, partial views, varying resolution, occlusion, and soft tissue deformation. Our novel body model has variable shape detail, allowing it to capture faces with a high-resolution deformable head model and body shape with lower-resolution. Finally we combine range data from an entire sequence to estimate a high-resolution displacement map that captures fine shape details. We compare our recovered models with high-resolution scans from a professional system and with avatars created by a commercial product. We extract accurate 3D avatars from challenging motion sequences and even capture soft tissue dynamics.
Similar papers:
  • Compositional Hierarchical Representation of Shape Manifolds for Classification of Non-Manifold Shapes [pdf] - Mete Ozay, Umit Rusen Aktas, Jeremy L. Wyatt, Ales Leonardis
  • The One Triangle Three Parallelograms Sampling Strategy and Its Application in Shape Regression [pdf] - Mikael Nilsson
  • Uncovering Interactions and Interactors: Joint Estimation of Head, Body Orientation and F-Formations From Surveillance Videos [pdf] - Elisa Ricci, Jagannadan Varadarajan, Ramanathan Subramanian, Samuel Rota Bulo, Narendra Ahuja, Oswald Lanz
  • Human Pose Estimation in Videos [pdf] - Dong Zhang, Mubarak Shah
  • BodyPrint: Pose Invariant 3D Shape Matching of Human Bodies [pdf] - Jiangping Wang, Kai Ma, Vivek Kumar Singh, Thomas Huang, Terrence Chen
Weakly-Supervised Alignment of Video With Text [pdf]
Piotr Bojanowski, Remi Lajugie, Edouard Grave, Francis Bach, Ivan Laptev, Jean Ponce, Cordelia Schmid

Abstract: Suppose that we are given a set of videos, along with natural language descriptions in the form of multiple sentences (e.g., manual annotations, movie scripts, sport summaries etc.), and that these sentences appear in the same temporal order as their visual counterparts. We propose in this paper a method for aligning the two modalities, i.e., automatically providing a time (frame) stamp for every sentence. Given vectorial features for both video and text, this can be cast as a temporal assignment problem, with an implicit linear mapping between the two feature modalities. We formulate this problem as an integer quadratic program, and solve its continuous convex relaxation using an efficient conditional gradient algorithm. Several rounding procedures are proposed to construct the final integer solution. After demonstrating significant improvements over the state of the art on the related task of aligning video with symbolic labels [7], we evaluate our method on a challenging dataset of videos with associated textual descriptions [37], and explore bagof-words and continuous representations for text.
Similar papers:
  • Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books [pdf] - Yukun Zhu, Ryan Kiros, Rich Zemel, Ruslan Salakhutdinov, Raquel Urtasun, Antonio Torralba, Sanja Fidler
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Semantic Video Entity Linking Based on Visual Content and Metadata [pdf] - Yuncheng Li, Xitong Yang, Jiebo Luo
  • Unsupervised Semantic Parsing of Video Collections [pdf] - Ozan Sener, Amir R. Zamir, Silvio Savarese, Ashutosh Saxena
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
A Groupwise Multilinear Correspondence Optimization for 3D Faces [pdf]
Timo Bolkart, Stefanie Wuhrer

Abstract: Multilinear face models are widely used to model the space of human faces with expressions. For databases of 3D human faces of different identities performing multiple expressions, these statistical shape models decouple identity and expression variations. To compute a high-quality multilinear face model, the quality of the registration of the database of 3D face scans used for training is essential. Meanwhile, a multilinear face model can be used as an effective prior to register 3D face scans, which are typically noisy and incomplete. Inspired by the minimum description length approach, we propose the first method to jointly optimize a multilinear model and the registration of the 3D scans used for training. Given an initial registration, our approach fully automatically improves the registration by optimizing an objective function that measures the compactness of the multilinear model, resulting in a sparse model. We choose a continuous representation for each face shape that allows to use a quasi-Newton method in parameter space for optimization. We show that our approach is computationally significantly more efficient and leads to correspondences of higher quality than existing methods based on linear statistical models. This allows us to evaluate our approach on large standard 3D face databases and in the presence of noisy initializations.
Similar papers:
  • PIEFA: Personalized Incremental and Ensemble Face Alignment [pdf] - Xi Peng, Shaoting Zhang, Yu Yang, Dimitris N. Metaxas
  • Learning to Predict Saliency on Face Images [pdf] - Mai Xu, Yun Ren, Zulin Wang
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • Shell PCA: Statistical Shape Modelling in Shell Space [pdf] - Chao Zhang, Behrend Heeren, Martin Rumpf, William A. P. Smith
Entropy-Based Latent Structured Output Prediction [pdf]
Diane Bouchacourt, Sebastian Nowozin, M. Pawan Kumar

Abstract: Recently several generalizations of the popular latent structural SVM framework have been proposed in the literature. Broadly speaking, the generalizations can be divided into two categories: (i) those that predict the output variables while either marginalizing the latent variables or estimating their most likely values; and (ii) those that predict the output variables by minimizing an entropy-based uncertainty measure over the latent space. In order to aid their application in computer vision, we study these generalizations with the aim of identifying their strengths and weaknesses. To this end, we propose a novel prediction criterion that includes as special cases all previous prediction criteria that have been used in the literature. Specifically, our frameworks prediction criterion minimizes the Aczel and Daroczy entropy of the output. This in turn allows us to design a learning objective that provides a unified framework (UF) for latent structured prediction. We develop a single optimization algorithm and empirically show that it is as effective as the more complex approaches that have been previously employed for latent structured prediction. Using this algorithm, we provide empirical evidence that lends support to prediction via the minimization of the latent space uncertainty.
Similar papers:
  • Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf] - Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic
  • Exploring Causal Relationships in Visual Object Tracking [pdf] - Karel Lebeda, Simon Hadfield, Richard Bowden
  • Similarity Gaussian Process Latent Variable Model for Multi-Modal Data Analysis [pdf] - Guoli Song, Shuhui Wang, Qingming Huang, Qi Tian
  • Learning Ensembles of Potential Functions for Structured Prediction With Latent Variables [pdf] - Hossein Hajimirsadeghi, Greg Mori
  • MANTRA: Minimum Maximum Latent Structural SVM for Image Classification and Ranking [pdf] - Thibaut Durand, Nicolas Thome, Matthieu Cord
Volumetric Bias in Segmentation and Reconstruction: Secrets and Solutions [pdf]
Yuri Boykov, Hossam Isack, Carl Olsson, Ismail Ben Ayed

Abstract: Many standard optimization methods for segmentation and reconstruction compute ML model estimates for appearance or geometry of segments, e.g. Zhu-Yuille [23], Torr [20], Chan-Vese [6], GrabCut [18], Delong et al. [8]. We observe that the standard likelihood term in these formulations corresponds to a generalized probabilistic K-means energy. In learning it is well known that this energy has a strong bias to clusters of equal size [11], which we express as a penalty for KL divergence from a uniform distribution of cardinalities. However, this volumetric bias has been mostly ignored in computer vision. We demonstrate significant artifacts in standard segmentation and reconstruction methods due to this bias. Moreover, we propose binary and multi-label optimization techniques that either (a) remove this bias or (b) replace it by a KL divergence term for any given target volume distribution. Our general ideas apply to continuous or discrete energy formulations in segmentation, stereo, and other reconstruction problems. Secrets (1) Solutions (6), (9-10) (a) GrabCut [18] with unbiased data term (10) (b) plane fitting [20, 8, 1] with unbiased data term (10)
Similar papers:
  • Higher-Order Inference for Multi-Class Log-Supermodular Models [pdf] - Jian Zhang, Josip Djolonga, Andreas Krause
  • A Multiscale Variable-Grouping Framework for MRF Energy Minimization [pdf] - Omer Meir, Meirav Galun, Stav Yagev, Ronen Basri, Irad Yavneh
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
  • Entropy Minimization for Convex Relaxation Approaches [pdf] - Mohamed Souiai, Martin R. Oswald, Youngwook Kee, Junmo Kim, Marc Pollefeys, Daniel Cremers
  • Secrets of GrabCut and Kernel K-Means [pdf] - Meng Tang, Ismail Ben Ayed, Dmitrii Marin, Yuri Boykov
Dense Semantic Correspondence Where Every Pixel is a Classifier [pdf]
Hilton Bristow, Jack Valmadre, Simon Lucey

Abstract: Determining dense semantic correspondences across objects and scenes is a difficult problem that underpins many higher-level computer vision algorithms. Unlike canonical dense correspondence problems which consider images that are spatially or temporally adjacent, semantic correspondence is characterized by images that share similar high-level structures whose exact appearance and geometry may differ. Motivated by object recognition literature and recent work on rapidly estimating linear classifiers, we treat semantic correspondence as a constrained detection problem, where an exemplar LDA classifier is learned for each pixel. LDA classifiers have two distinct benefits: (i) they exhibit higher average precision than similarity metrics typically used in correspondence problems, and (ii) unlike exemplar SVM, can output globally interpretable posterior probabilities without calibration, whilst also being significantly faster to train. We pose the correspondence problem as a graphical model, where the unary potentials are computed via convolution with the set of exemplar classifiers, and the joint potentials enforce smoothly varying correspondence assignment.
Similar papers:
  • Oriented Light-Field Windows for Scene Flow [pdf] - Pratul P. Srinivasan, Michael W. Tao, Ren Ng, Ravi Ramamoorthi
  • Person Re-Identification With Correspondence Structure Learning [pdf] - Yang Shen, Weiyao Lin, Junchi Yan, Mingliang Xu, Jianxin Wu, Jingdong Wang
  • Face Flow [pdf] - Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou
  • Large Displacement 3D Scene Flow With Occlusion Reasoning [pdf] - Andrei Zanfir, Cristian Sminchisescu
  • Flow Fields: Dense Correspondence Fields for Highly Accurate Large Displacement Optical Flow Estimation [pdf] - Christian Bailer, Bertram Taetz, Didier Stricker
Globally Optimal 2D-3D Registration From Points or Lines Without Correspondences [pdf]
Mark Brown, David Windridge, Jean-Yves Guillemaut

Abstract: We present a novel approach to 2D-3D registration from points or lines without correspondences. While there exist established solutions in the case where correspondences are known, there are many situations where it is not possible to reliably extract such correspondences across modalities, thus requiring the use of a correspondence-free registration algorithm. Existing correspondence-free methods rely on local search strategies and consequently have no guarantee of finding the optimal solution. In contrast, we present the first globally optimal approach to 2D-3D registration without correspondences, achieved by a Branch-and-Bound algorithm. Furthermore, a deterministic annealing procedure is proposed to speed up the nested branch-and-bound algorithm used. The theoretical and practical advantages this brings are demonstrated on a range of synthetic and real data where it is observed that the proposed approach is significantly more robust to high proportions of outliers compared to existing approaches.
Similar papers:
  • Camera Pose Voting for Large-Scale Image-Based Localization [pdf] - Bernhard Zeisl, Torsten Sattler, Marc Pollefeys
  • An Adaptive Data Representation for Robust Point-Set Registration and Merging [pdf] - Dylan Campbell, Lars Petersson
  • An Efficient Minimal Solution for Multi-Camera Motion [pdf] - Jonathan Ventura, Clemens Arth, Vincent Lepetit
  • Guaranteed Outlier Removal for Rotation Search [pdf] - Alvaro Parra Bustos, Tat-Jun Chin
  • Robust and Optimal Sum-of-Squares-Based Point-to-Plane Registration of Image Sets and Structured Scenes [pdf] - Danda Pani Paudel, Adlane Habed, Cedric Demonceaux, Pascal Vasseur
FASText: Efficient Unconstrained Scene Text Detector [pdf]
Michal Busta, Lukas Neumann, Jiri Matas

Abstract: We propose a novel easy-to-implement stroke detector based on an efcient pixel intensity comparison to surrounding pixels. Stroke-specic keypoints are efciently detected and text fragments are subsequently extracted by local thresholding guided by keypoint properties. Classication based on effectively calculated features then eliminates non-text regions. The stroke-specic keypoints produce 2 times less region segmentations and still detect 25% more characters than the commonly exploited MSER detector and the process is 4 times faster. After a novel efcient classication step, the number of regions is reduced to 7 times less than the standard method and is still almost 3 times faster. All stages of the proposed pipeline are scale- and rotation-invariant and support a wide variety of scripts (Latin, Hebrew, Chinese, etc.) and fonts. When the proposed detector is plugged into a scene text localization and recognition pipeline, a state-of-the-art text localization accuracy is maintained whilst the processing time is signicantly reduced.
Similar papers:
  • Extraction of Virtual Baselines From Distorted Document Images Using Curvilinear Projection [pdf] - Gaofeng Meng, Zuming Huang, Yonghong Song, Shiming Xiang, Chunhong Pan
  • Learning Common Sense Through Visual Abstraction [pdf] - Ramakrishna Vedantam, Xiao Lin, Tanmay Batra, C. Lawrence Zitnick, Devi Parikh
  • Predicting Deep Zero-Shot Convolutional Neural Networks Using Textual Descriptions [pdf] - Jimmy Lei Ba, Kevin Swersky, Sanja Fidler, Ruslan salakhutdinov
  • Lost Shopping! Monocular Localization in Large Indoor Spaces [pdf] - Shenlong Wang, Sanja Fidler, Raquel Urtasun
  • Text Flow: A Unified Text Detection System in Natural Scene Images [pdf] - Shangxuan Tian, Yifeng Pan, Chang Huang, Shijian Lu, Kai Yu, Chew Lim Tan
Guaranteed Outlier Removal for Rotation Search [pdf]
Alvaro Parra Bustos, Tat-Jun Chin

Abstract: Rotation search has become a core routine for solving many computer vision problems. The aim is to rotationally align two input point sets with correspondences. Recently, there is significant interest in developing globally optimal rotation search algorithms. A notable weakness of global algorithms, however, is their relatively high computational cost, especially on large problem sizes and data with a high proportion of outliers. In this paper, we propose a novel outlier removal technique for rotation search. Our method guarantees that any correspondence it discards as an outlier does not exist in the inlier set of the globally optimal rotation for the original data. Based on simple geometric operations, our algorithm is deterministic and fast. Used as a preprocessor to prune a large portion of the outliers from the input data, our method enables substantial speed-up of rotation search algorithms without compromising global optimality. We demonstrate the efficacy of our method in various synthetic and real data experiments1 .
Similar papers:
  • Camera Pose Voting for Large-Scale Image-Based Localization [pdf] - Bernhard Zeisl, Torsten Sattler, Marc Pollefeys
  • An Adaptive Data Representation for Robust Point-Set Registration and Merging [pdf] - Dylan Campbell, Lars Petersson
  • An Efficient Minimal Solution for Multi-Camera Motion [pdf] - Jonathan Ventura, Clemens Arth, Vincent Lepetit
  • Learning Discriminative Reconstructions for Unsupervised Outlier Removal [pdf] - Yan Xia, Xudong Cao, Fang Wen, Gang Hua, Jian Sun
  • Globally Optimal 2D-3D Registration From Points or Lines Without Correspondences [pdf] - Mark Brown, David Windridge, Jean-Yves Guillemaut
Semantically-Aware Aerial Reconstruction From Multi-Modal Data [pdf]
Randi Cabezas, Julian Straub, John W. Fisher III

Abstract: We consider a methodology for integrating multiple sensors along with semantic information to enhance scene representations. We propose a probabilistic generative model for inferring semantically-informed aerial reconstructions from multi-modal data within a consistent mathematical framework. The approach, called SemanticallyAware Aerial Reconstruction (SAAR), not only exploits inferred scene geometry, appearance, and semantic observations to obtain a meaningful categorization of the data, but also extends previously proposed methods by imposing structure on the prior over geometry, appearance, and semantic labels. This leads to more accurate reconstructions and the ability to fill in missing contextual labels via joint sensor and semantic information. We introduce a new multi-modal synthetic dataset in order to provide quantitative performance analysis. Additionally, we apply the model to real-world data and exploit OpenStreetMap as a source of semantic observations. We show quantitative improvements in reconstruction accuracy of large-scale urban scenes from the combination of LiDAR, aerial photography, and semantic data. Furthermore, we demonstrate the models ability to fill in for missing sensed data, leading to more interpretable reconstructions.
Similar papers:
  • Segment-Phrase Table for Semantic Segmentation, Visual Entailment and Paraphrasing [pdf] - Hamid Izadinia, Fereshteh Sadeghi, Santosh K. Divvala, Hannaneh Hajishirzi, Yejin Choi, Ali Farhadi
  • Conditional Random Fields as Recurrent Neural Networks [pdf] - Shuai Zheng, Sadeep Jayasumana, Bernardino Romera-Paredes, Vibhav Vineet, Zhizhong Su, Dalong Du, Chang Huang, Philip H. S. Torr
  • Semantic Component Analysis [pdf] - Calvin Murdock, Fernando De la Torre
  • Joint Object and Part Segmentation Using Deep Learned Potentials [pdf] - Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille
  • ML-MG: Multi-Label Learning With Missing Labels Using a Mixed Graph [pdf] - Baoyuan Wu, Siwei Lyu, Bernard Ghanem
Learning Complexity-Aware Cascades for Deep Pedestrian Detection [pdf]
Zhaowei Cai, Mohammad Saberian, Nuno Vasconcelos

Abstract: The design of complexity-aware cascaded detectors, combining features of very different complexities, is considered. A new cascade design procedure is introduced, by formulating cascade learning as the Lagrangian optimization of a risk that accounts for both accuracy and complexity. A boosting algorithm, denoted as complexity aware cascade training (CompACT), is then derived to solve this optimization. CompACT cascades are shown to seek an optimal trade-off between accuracy and complexity by pushing features of higher complexity to the later cascade stages, where only a few difficult candidate patches remain to be classified. This enables the use of features of vastly different complexities in a single detector. In result, the feature pool can be expanded to features previously impractical for cascade design, such as the responses of a deep convolutional neural network (CNN). This is demonstrated through the design of a pedestrian detector with a pool of features whose complexities span orders of magnitude. The resulting cascade generalizes the combination of a CNN with an object proposal mechanism: rather than a pre-processing stage, CompACT cascades seamlessly integrate CNNs in their stages. This enables state of the art performance on the Caltech and KITTI datasets, at fairly fast speeds.
Similar papers:
  • Deep Networks for Image Super-Resolution With Sparse Prior [pdf] - Zhaowen Wang, Ding Liu, Jianchao Yang, Wei Han, Thomas Huang
  • Robust Optimization for Deep Regression [pdf] - Vasileios Belagiannis, Christian Rupprecht, Gustavo Carneiro, Nassir Navab
  • Deep Learning Strong Parts for Pedestrian Detection [pdf] - Yonglong Tian, Ping Luo, Xiaogang Wang, Xiaoou Tang
  • Convolutional Channel Features [pdf] - Bin Yang, Junjie Yan, Zhen Lei, Stan Z. Li
  • Object Detection Using Generalization and Efficiency Balanced Co-Occurrence Features [pdf] - Haoyu Ren, Ze-Nian Li
Active Object Localization With Deep Reinforcement Learning [pdf]
Juan C. Caicedo, Svetlana Lazebnik

Abstract: We present an active detection model for localizing objects in scenes. The model is class-specific and allows an agent to focus attention on candidate regions for identifying the correct location of a target object. This agent learns to deform a bounding box using simple transformation actions, with the goal of determining the most specific location of target objects following top-down reasoning. The proposed localization agent is trained using deep reinforcement learning, and evaluated on the Pascal VOC 2007 dataset. We show that agents guided by the proposed model are able to localize a single instance of an object after analyzing only between 11 and 25 regions in an image, and obtain the best detection results among systems that do not use object proposals for object localization.
Similar papers:
  • Object Detection via a Multi-Region and Semantic Segmentation-Aware CNN Model [pdf] - Spyros Gidaris, Nikos Komodakis
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Contextual Action Recognition With R*CNN [pdf] - Georgia Gkioxari, Ross Girshick, Jitendra Malik
  • Learning to Track for Spatio-Temporal Action Localization [pdf] - Philippe Weinzaepfel, Zaid Harchaoui, Cordelia Schmid
Adaptive Hashing for Fast Similarity Search [pdf]
Fatih Cakir, Stan Sclaroff

Abstract: With the staggering growth in image and video datasets, algorithms that provide fast similarity search and compact storage are crucial. Hashing methods that map the data into Hamming space have shown promise; however, many of these methods employ a batch-learning strategy in which the computational cost and memory requirements may become intractable and infeasible with larger and larger datasets. To overcome these challenges, we propose an online learning algorithm based on stochastic gradient descent in which the hash functions are updated iteratively with streaming data. In experiments with three image retrieval benchmarks, our online algorithm attains retrieval accuracy that is comparable to competing state-of-the-art batch-learning solutions, while our formulation is orders of magnitude faster and being online it is adaptable to the variations of the data. Moreover, our formulation yields improved retrieval performance over a recently reported online hashing technique, Online Kernel Hashing.
Similar papers:
  • An NMF Perspective on Binary Hashing [pdf] - Lopamudra Mukherjee, Sathya N. Ravi, Vamsi K. Ithapu, Tyler Holmes, Vikas Singh
  • Learning Binary Codes for Maximum Inner Product Search [pdf] - Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen
  • Alternating Co-Quantization for Cross-Modal Hashing [pdf] - Go Irie, Hiroyuki Arai, Yukinobu Taniguchi
  • Multi-View Complementary Hash Tables for Nearest Neighbor Search [pdf] - Xianglong Liu, Lei Huang, Cheng Deng, Jiwen Lu, Bo Lang
  • kNN Hashing With Factorized Neighborhood Representation [pdf] - Kun Ding, Chunlei Huo, Bin Fan, Chunhong Pan
An Adaptive Data Representation for Robust Point-Set Registration and Merging [pdf]
Dylan Campbell, Lars Petersson

Abstract: This paper presents a framework for rigid point-set registration and merging using a robust continuous data representation. Our point-set representation is constructed by training a one-class support vector machine with a Gaussian radial basis function kernel and subsequently approximating the output function with a Gaussian mixture model. We leverage the representations sparse parametrisation and robustness to noise, outliers and occlusions in an efficient registration algorithm that minimises the L2 distance between our support vectorparametrised Gaussian mixtures. In contrast, existing techniques, such as Iterative Closest Point and Gaussian mixture approaches, manifest a narrower region of convergence and are less robust to occlusions and missing data, as demonstrated in the evaluation on a range of 2D and 3D datasets. Finally, we present a novel algorithm, GMMerge, that parsimoniously and equitably merges aligned mixture models, allowing the framework to be used for reconstruction and mapping.
Similar papers:
  • Robust and Optimal Sum-of-Squares-Based Point-to-Plane Registration of Image Sets and Structured Scenes [pdf] - Danda Pani Paudel, Adlane Habed, Cedric Demonceaux, Pascal Vasseur
  • Registering Images to Untextured Geometry Using Average Shading Gradients [pdf] - Tobias Plotz, Stefan Roth
  • Globally Optimal 2D-3D Registration From Points or Lines Without Correspondences [pdf] - Mark Brown, David Windridge, Jean-Yves Guillemaut
  • Guaranteed Outlier Removal for Rotation Search [pdf] - Alvaro Parra Bustos, Tat-Jun Chin
  • Robust Nonrigid Registration by Convex Optimization [pdf] - Qifeng Chen, Vladlen Koltun
Non-Parametric Structure-Based Calibration of Radially Symmetric Cameras [pdf]
Federico Camposeco, Torsten Sattler, Marc Pollefeys

Abstract: We propose a novel two-step method for estimating the intrinsic and extrinsic calibration of any radially symmetric camera, including non-central systems. The first step consists of estimating the camera pose, given a Structure from Motion (SfM) model, up to the translation along the optical axis. As a second step, we obtain the calibration by finding the translation of the camera center using an ordering constraint. The method makes use of the 1D radial camera model, which allows us to effectively handle any radially symmetric camera, including non-central ones. Using this ordering constraint, we show that the we are able to calibrate several different (central and non-central) Wide Field of View (WFOV) cameras, including fisheye, hypercatadioptric and spherical catadioptric cameras, as well as pinhole cameras, using a single image or jointly solving for several views.
Similar papers:
  • Structure From Motion Using Structure-Less Resection [pdf] - Enliang Zheng, Changchang Wu
  • Efficient Solution to the Epipolar Geometry for Radially Distorted Cameras [pdf] - Zuzana Kukelova, Jan Heller, Martin Bujnak, Andrew Fitzgibbon, Tomas Pajdla
  • Accurate Camera Calibration Robust to Defocus Using a Smartphone [pdf] - Hyowon Ha, Yunsu Bok, Kyungdon Joo, Jiyoung Jung, In So Kweon
  • A Linear Generalized Camera Calibration From Three Intersecting Reference Planes [pdf] - Mai Nishimura, Shohei Nobuhara, Takashi Matsuyama, Shinya Shimizu, Kensaku Fujii
  • On the Equivalence of Moving Entrance Pupil and Radial Distortion for Camera Calibration [pdf] - Avinash Kumar, Narendra Ahuja
A Wavefront Marching Method for Solving the Eikonal Equation on Cartesian Grids [pdf]
Brais Cancela, Marcos Ortega, Manuel G. Penedo

Abstract: This paper presents a new wavefront propagation method for dealing with the classic Eikonal equation. While classic Dijkstra-like graph-based techniques achieve the solution in O(M log M ), they do not approximate the unique physically relevant solution very well. Fast Marching Methods (FMM) were created to efficiently solve the continuous problem. The proposed approximation tries to maintain the complexity, in order to make the algorithm useful in a wide range of contexts. The key idea behind our method is the creation of mini wave-fronts, which are combined to propagate the solution. Experimental results show the improvement in the accuracy with respect to the state of the art, while the average computational speed is maintained in O(M log M ), similar to the FMM techniques.
Similar papers:
  • Context-Guided Diffusion for Label Propagation on Graphs [pdf] - Kwang In Kim, James Tompkin, Hanspeter Pfister, Christian Theobalt
  • Convex Optimization With Abstract Linear Operators [pdf] - Steven Diamond, Stephen Boyd
  • Classical Scaling Revisited [pdf] - Gil Shamai, Yonathan Aflalo, Michael Zibulevsky, Ron Kimmel
  • Fast and Effective L0 Gradient Minimization by Region Fusion [pdf] - Rang M. H. Nguyen, Michael S. Brown
  • Interpolation on the Manifold of K Component GMMs [pdf] - Hyunwoo J. Kim, Nagesh Adluru, Monami Banerjee, Baba C. Vemuri, Vikas Singh
Look and Think Twice: Capturing Top-Down Visual Attention With Feedback Convolutional Neural Networks [pdf]
Chunshui Cao, Xianming Liu, Yi Yang, Yinan Yu, Jiang Wang, Zilei Wang, Yongzhen Huang, Liang Wang, Chang Huang, Wei Xu, Deva Ramanan, Thomas S. Huang

Abstract: While feedforward deep convolutional neural networks (CNNs) have been a great success in computer vision, it is important to note that the human visual cortex generally contains more feedback than feedforward connections. In this paper, we will briefly introduce the background of feedbacks in the human visual cortex, which motivates us to develop a computational feedback mechanism in deep neural networks. In addition to the feedforward inference in traditional neural networks, a feedback loop is introduced to infer the activation status of hidden layer neurons according to the goal of the network, e.g., high-level semantic labels. We analogize this mechanism as Look and Think Twice. The feedback networks help better visualize and understand how deep neural networks work, and capture visual attention on expected objects, even in images with cluttered background and multiple objects. Experiments on ImageNet dataset demonstrate its effectiveness in solving tasks such as image classification and object localization.
Similar papers:
  • HARF: Hierarchy-Associated Rich Features for Salient Object Detection [pdf] - Wenbin Zou, Nikos Komodakis
  • Object Detection via a Multi-Region and Semantic Segmentation-Aware CNN Model [pdf] - Spyros Gidaris, Nikos Komodakis
  • Training a Feedback Loop for Hand Pose Estimation [pdf] - Markus Oberweger, Paul Wohlhart, Vincent Lepetit
  • Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification [pdf] - Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun
  • Relaxing From Vocabulary: Robust Weakly-Supervised Deep Learning for Vocabulary-Free Image Tagging [pdf] - Jianlong Fu, Yue Wu, Tao Mei, Jinqiao Wang, Hanqing Lu, Yong Rui
Low-Rank Matrix Factorization Under General Mixture Noise Distributions [pdf]
Xiangyong Cao, Yang Chen, Qian Zhao, Deyu Meng, Yao Wang, Dong Wang, Zongben Xu

Abstract: Error 1 4 2 x 10 1.5 Many computer vision problems can be posed as learning a low-dimensional subspace from high dimensional data. The low rank matrix factorization (LRMF) represents a commonly utilized subspace learning strategy. Most of the current LRMF techniques are constructed on the optimization problem using L1 norm and L2 norm, which mainly deal with Laplacian and Gaussian noise, respectively. To make LRMF capable of adapting more complex noise, this paper proposes a new LRMF model by assuming noise as Mixture of Exponential Power (MoEP) distributions and proposes a penalized MoEP model by combining the penalized likelihood method with MoEP distributions. Such setting facilitates the learned LRMF model capable of automatically fitting the real noise through MoEP distributions. Each component in this mixture is adapted from a series of preliminary super- or sub-Gaussian candidates. An Expectation Maximization (EM) algorithm is also designed to infer the parameters involved in the proposed PMoEP model. The advantage of our method is demonstrated by extensive experiments on synthetic data, face modeling and hyperspectral image restoration.
Similar papers:
  • Hyperspectral Compressive Sensing Using Manifold-Structured Sparsity Prior [pdf] - Lei Zhang, Wei Wei, Yanning Zhang, Fei Li, Chunhua Shen, Qinfeng Shi
  • Secrets of Matrix Factorization: Approximations, Numerics, Manifold Optimization and Random Restarts [pdf] - Je Hyeong Hong, Andrew Fitzgibbon
  • The Likelihood-Ratio Test and Efficient Robust Estimation [pdf] - Andrea Cohen, Christopher Zach
  • Shape Interaction Matrix Revisited and Robustified: Efficient Subspace Clustering With Corrupted and Incomplete Data [pdf] - Pan Ji, Mathieu Salzmann, Hongdong Li
  • An Efficient Statistical Method for Image Noise Level Estimation [pdf] - Guangyong Chen, Fengyuan Zhu, Pheng Ann Heng
Weakly-Supervised Structured Output Learning With Flexible and Latent Graphs Using High-Order Loss Functions [pdf]
Gustavo Carneiro, Tingying Peng, Christine Bayer, Nassir Navab

Abstract: We introduce two new structured output models that use a latent graph, which is flexible in terms of the number of nodes and structure, where the training process minimises a high-order loss function using a weakly annotated training set. These models are developed in the context of microscopy imaging of malignant tumours, where the estimation of the number and proportion of classes of microcirculatory supply units (MCSU) is important in the assessment of the efficacy of common cancer treatments (an MCSU is a region of the tumour tissue supplied by a microvessel). The proposed methodologies take as input multimodal microscopy images of a tumour, and estimate the number and proportion of MCSU classes. This estimation is facilitated by the use of an underlying latent graph (not present in the manual annotations), where each MCSU is represented by a node in this graph, labelled with the MCSU class and image location. The training process uses the manual weak annotations available, consisting of the number of MCSU classes per training image, where the training objective is the minimisation of a high-order loss function based on the norm of the error between the manual and estimated annotations. One of the models proposed is based on a new flexible latent structure support vector machine (FLSSVM) and the other is based on a deep convolutional neural network (DCNN) model. Using a dataset of 89 weakly annotated pairs of multimodal images from eight tumours, we show that the quantita
Similar papers:
  • Similarity Gaussian Process Latent Variable Model for Multi-Modal Data Analysis [pdf] - Guoli Song, Shuhui Wang, Qingming Huang, Qi Tian
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • MANTRA: Minimum Maximum Latent Structural SVM for Image Classification and Ranking [pdf] - Thibaut Durand, Nicolas Thome, Matthieu Cord
  • Cutting Edge: Soft Correspondences in Multimodal Scene Parsing [pdf] - Sarah Taghavi Namin, Mohammad Najafi, Mathieu Salzmann, Lars Petersson
  • Weakly- and Semi-Supervised Learning of a Deep Convolutional Network for Semantic Image Segmentation [pdf] - George Papandreou, Liang-Chieh Chen, Kevin P. Murphy, Alan L. Yuille
Recursive Frechet Mean Computation on the Grassmannian and its Applications to Computer Vision [pdf]
Rudrasis Chakraborty, Baba C. Vemuri

Abstract: In the past decade, Grassmann manifolds (Grassmannian) have been commonly used in mathematical formulations of many Computer Vision tasks. Averaging points on a Grassmann manifold is a very common operation in many applications including but not limited to, tracking, action recognition, video-face recognition, face recognition, etc. Computing the intrinsic/Frechet mean (FM) of a set of points on the Grassmann can be cast as finding the global optimum (if it exists) of the sum of squared geodesic distances cost function. A common approach to solve this problem involves the use of the gradient descent method. An alternative way to compute the FM is to develop a recursive/inductive definition that does not involve optimizing the aforementioned cost function. In this paper, we propose one such computationally efficient algorithm called the Grassmann inductive Frechet mean estimator (GiFME). In developing the recursive solution to find the FM of the given set of points, GiFME exploits the fact that there is a closed form solution to find the FM of two points on the Grassmann. In the limit as the number of samples tends to infinity, we prove that GiFME converges to the FM (this is called the weak consistency result on the Grassmann manifold). Further, for the finite sample case, in the limit as the number of sample paths (trials) goes to infinity, we show that GiFME converges to the finite sample FM. Moreover, we present a bound on the geodesic distance between the estimate from
Similar papers:
  • Local Subspace Collaborative Tracking [pdf] - Lin Ma, Xiaoqin Zhang, Weiming Hu, Junliang Xing, Jiwen Lu, Jie Zhou
  • Highly-Expressive Spaces of Well-Behaved Transformations: Keeping It Simple [pdf] - Oren Freifeld, Soren Hauberg, Kayhan Batmanghelich, John W. Fisher III
  • Shell PCA: Statistical Shape Modelling in Shell Space [pdf] - Chao Zhang, Behrend Heeren, Martin Rumpf, William A. P. Smith
  • Classical Scaling Revisited [pdf] - Gil Shamai, Yonathan Aflalo, Michael Zibulevsky, Ron Kimmel
  • Beyond Gauss: Image-Set Matching on the Riemannian Manifold of PDFs [pdf] - Mehrtash Harandi, Mathieu Salzmann, Mahsa Baktashmotlagh
HICO: A Benchmark for Recognizing Human-Object Interactions in Images [pdf]
Yu-Wei Chao, Zhan Wang, Yugeng He, Jiaxuan Wang, Jia Deng

Abstract: We introduce a new benchmark Humans Interacting with Common Objects (HICO) for recognizing humanobject interactions (HOI). We demonstrate the key features of HICO: a diverse set of interactions with common object categories, a list of well-defined, sense-based HOI categories, and an exhaustive labeling of co-occurring interactions with an object category in each image. We perform an in-depth analysis of representative current approaches and show that DNNs enjoy a significant edge. In addition, we show that semantic knowledge can significantly improve HOI recognition, especially for uncommon categories.
Similar papers:
  • Action Localization in Videos Through Context Walk [pdf] - Khurram Soomro, Haroon Idrees, Mubarak Shah
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
A Nonparametric Bayesian Approach Toward Stacked Convolutional Independent Component Analysis [pdf]
Sotirios P. Chatzis, Dimitrios Kosmopoulos

Abstract: Unsupervised feature learning algorithms based on convolutional formulations of independent components analysis (ICA) have been demonstrated to yield state-of-the-art results in several action recognition benchmarks. However, existing approaches do not allow for the number of latent components (features) to be automatically inferred from the data in an unsupervised manner. This is a significant disadvantage of the state-of-the-art, as it results in considerable burden imposed on researchers and practitioners, who must resort to tedious cross-validation procedures to obtain the optimal number of latent features. To resolve these issues, in this paper we introduce a convolutional nonparametric Bayesian sparse ICA architecture for overcomplete feature learning from high-dimensional data. Our method utilizes an Indian buffet process prior to facilitate inference of the appropriate number of latent features under a hybrid variational inference algorithm, scalable to massive datasets. As we show, our model can be naturally used to obtain deep unsupervised hierarchical feature extractors, by greedily stacking successive model layers, similar to existing approaches. In addition, inference for this model is completely heuristics-free; thus, it obviates the need of tedious parameter tuning, which is a major challenge most deep learning approaches are faced with. We evaluate our method on several action recognition benchmarks, and exhibit its advantages over the state-of-the-art.
Similar papers:
  • Entropy-Based Latent Structured Output Prediction [pdf] - Diane Bouchacourt, Sebastian Nowozin, M. Pawan Kumar
  • MANTRA: Minimum Maximum Latent Structural SVM for Image Classification and Ranking [pdf] - Thibaut Durand, Nicolas Thome, Matthieu Cord
  • Cutting Edge: Soft Correspondences in Multimodal Scene Parsing [pdf] - Sarah Taghavi Namin, Mohammad Najafi, Mathieu Salzmann, Lars Petersson
  • Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf] - Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic
  • Similarity Gaussian Process Latent Variable Model for Multi-Modal Data Analysis [pdf] - Guoli Song, Shuhui Wang, Qingming Huang, Qi Tian
A Deep Visual Correspondence Embedding Model for Stereo Matching Costs [pdf]
Zhuoyuan Chen, Xun Sun, Liang Wang, Yinan Yu, Chang Huang

Abstract: This paper presents a data-driven matching cost for stereo matching. A novel deep visual correspondence embedding model is trained via Convolutional Neural Network on a large set of stereo images with ground truth disparities. This deep embedding model leverages appearance data to learn visual similarity relationships between corresponding image patches, and explicitly maps intensity values into an embedding feature space to measure pixel dissimilarities. Experimental results on KITTI and Middlebury data sets demonstrate the effectiveness of our model. First, we prove that the new measure of pixel dissimilarity outperforms traditional matching costs. Furthermore, when integrated with a global stereo framework, our method ranks top 3 among all two-frame algorithms on the KITTI benchmark. Finally, cross-validation results show that our model is able to make correct predictions for unseen data which are outside of its labeled training set. (a) Reference image of a stereo pair (b) Winner-takes-all result (c) Result from an MRF-based stereo framework
Similar papers:
  • Blur-Aware Disparity Estimation From Defocus Stereo Images [pdf] - Ching-Hui Chen, Hui Zhou, Timo Ahonen
  • StereoSnakes: Contour Based Consistent Object Extraction For Stereo Images [pdf] - Ran Ju, Tongwei Ren, Gangshan Wu
  • The HCI Stereo Metrics: Geometry-Aware Performance Analysis of Stereo Algorithms [pdf] - Katrin Honauer, Lena Maier-Hein, Daniel Kondermann
  • MAP Disparity Estimation Using Hidden Markov Trees [pdf] - Eric T. Psota, Jedrzej Kowalczuk, Mateusz Mittek, Lance C. Perez
  • Depth Map Estimation and Colorization of Anaglyph Images Using Local Color Prior and Reverse Intensity Distribution [pdf] - Williem, Ramesh Raskar, In Kyu Park
Action Detection by Implicit Intentional Motion Clustering [pdf]
Wei Chen, Jason J. Corso

Abstract: Explicitly using human detection and pose estimation has found limited success in action recognition problems. This may be due to the complexity in the articulated motion human exhibit. Yet, we know that action requires an actor and intention. This paper hence seeks to understand the spatiotemporal properties of intentional movement and how to capture such intentional movement without relying on challenging human detection and tracking. We conduct a quantitative analysis of intentional movement, and our findings motivate a new approach for implicit intentional movement extraction that is based on spatiotemporal trajectory clustering by leveraging the properties of intentional movement. The intentional movement clusters are then used as action proposals for detection. Our results on three action detection benchmarks indicate the relevance of focusing on intentional movement for action detection; our method significantly outperforms the state of the art on the challenging MSR-II multi-action video benchmark.
Similar papers:
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Action Recognition by Hierarchical Mid-Level Action Elements [pdf] - Tian Lan, Yuke Zhu, Amir Roshan Zamir, Silvio Savarese
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
An Efficient Statistical Method for Image Noise Level Estimation [pdf]
Guangyong Chen, Fengyuan Zhu, Pheng Ann Heng

Abstract: In this paper, we address the problem of estimating noise level from a single image contaminated by additive zeromean Gaussian noise. We first provide rigorous analysis on the statistical relationship between the noise variance and the eigenvalues of the covariance matrix of patches within an image, which shows that many state-of-the-art noise estimation methods underestimate the noise level of an image. To this end, we derive a new nonparametric algorithm for efficient noise level estimation based on the observation that patches decomposed from a clean image often lie around a low-dimensional subspace. The performance of our method has been guaranteed both theoretically and empirically. Specifically, our method outperforms existing state-of-theart algorithms on estimating noise level with the least executing time in our experiments. We further demonstrate that the denoising algorithm BM3D algorithm achieves optimal performance using noise variance estimated by our algorithm.
Similar papers:
  • Learning Nonlinear Spectral Filters for Color Image Reconstruction [pdf] - Michael Moeller, Julia Diebold, Guy Gilboa, Daniel Cremers
  • The Likelihood-Ratio Test and Efficient Robust Estimation [pdf] - Andrea Cohen, Christopher Zach
  • Patch Group Based Nonlocal Self-Similarity Prior Learning for Image Denoising [pdf] - Jun Xu, Lei Zhang, Wangmeng Zuo, David Zhang, Xiangchu Feng
  • External Patch Prior Guided Internal Clustering for Image Denoising [pdf] - Fei Chen, Lei Zhang, Huimin Yu
  • Low-Rank Matrix Factorization Under General Mixture Noise Distributions [pdf] - Xiangyong Cao, Yang Chen, Qian Zhao, Deyu Meng, Yao Wang, Dong Wang, Zongben Xu
Blur-Aware Disparity Estimation From Defocus Stereo Images [pdf]
Ching-Hui Chen, Hui Zhou, Timo Ahonen

Abstract: Defocus blur usually causes performance degradation in establishing the visual correspondence between stereo images. We propose a blur-aware disparity estimation method that is robust to the mismatch of focus in stereo images. The relative blur resulting from the mismatch of focus between stereo images is approximated as the difference of the square diameters of the blur kernels. Based on the defocus and stereo model, we propose the relative blur versus disparity (RBD) model that characterizes the relative blur as a second-order polynomial function of disparity. Our method alternates between RBD model update and disparity update in each iteration. The RBD model in return refines the disparity estimation by updating the matching cost and aggregation weight to compensate the mismatch of focus. Experiments using both synthesized and real datasets demonstrate the effectiveness of our proposed algorithm.
Similar papers:
  • The HCI Stereo Metrics: Geometry-Aware Performance Analysis of Stereo Algorithms [pdf] - Katrin Honauer, Lena Maier-Hein, Daniel Kondermann
  • Depth Map Estimation and Colorization of Anaglyph Images Using Local Color Prior and Reverse Intensity Distribution [pdf] - Williem, Ramesh Raskar, In Kyu Park
  • Depth Recovery From Light Field Using Focal Stack Symmetry [pdf] - Haiting Lin, Can Chen, Sing Bing Kang, Jingyi Yu
  • A Deep Visual Correspondence Embedding Model for Stereo Matching Costs [pdf] - Zhuoyuan Chen, Xun Sun, Liang Wang, Yinan Yu, Chang Huang
  • MAP Disparity Estimation Using Hidden Markov Trees [pdf] - Eric T. Psota, Jedrzej Kowalczuk, Mateusz Mittek, Lance C. Perez
DeepDriving: Learning Affordance for Direct Perception in Autonomous Driving [pdf]
Chenyi Chen, Ari Seff, Alain Kornhauser, Jianxiong Xiao

Abstract: Today, there are two major paradigms for vision-based autonomous driving systems: mediated perception approaches that parse an entire scene to make a driving decision, and behavior reflex approaches that directly map an input image to a driving action by a regressor. In this paper, we propose a third paradigm: a direct perception approach to estimate the affordance for driving. We propose to map an input image to a small number of key perception indicators that directly relate to the affordance of a road/traffic state for driving. Our representation provides a set of compact yet complete descriptions of the scene to enable a simple controller to drive autonomously. Falling in between the two extremes of mediated perception and behavior reflex, we argue that our direct perception representation provides the right level of abstraction. To demonstrate this, we train a deep Convolutional Neural Network using recording from 12 hours of human driving in a video game and show that our model can work well to drive a car in a very diverse set of virtual environments. We also train a model for car distance estimation on the KITTI dataset. Results show that our direct perception approach can generalize well to real driving images. Source code and data are available on our project website.
Similar papers:
  • Learning to See by Moving [pdf] - Pulkit Agrawal, Joao Carreira, Jitendra Malik
  • Deep Learning Strong Parts for Pedestrian Detection [pdf] - Yonglong Tian, Ping Luo, Xiaogang Wang, Xiaoou Tang
  • Robust Optimization for Deep Regression [pdf] - Vasileios Belagiannis, Christian Rupprecht, Gustavo Carneiro, Nassir Navab
  • Real-Time Pose Estimation Piggybacked on Object Detection [pdf] - Roman Juranek, Adam Herout, Marketa Dubska, Pavel Zemcik
  • PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization [pdf] - Alex Kendall, Matthew Grimes, Roberto Cipolla
External Patch Prior Guided Internal Clustering for Image Denoising [pdf]
Fei Chen, Lei Zhang, Huimin Yu

Abstract: Natural image modeling plays a key role in many vision problems such as image denoising. Image priors are widely used to regularize the denoising process, which is an illposed inverse problem. One category of denoising methods exploit the priors (e.g., TV, sparsity) learned from external clean images to reconstruct the given noisy image, while another category of methods exploit the internal prior (e.g., self-similarity) to reconstruct the latent image. Though the internal prior based methods have achieved impressive denoising results, the improvement of visual quality will become very difficult with the increase of noise level. In this paper, we propose to exploit image external patch prior and internal self-similarity prior jointly, and develop an external patch prior guided internal clustering algorithm for image denoising. It is known that natural image patches form multiple subspaces. By utilizing Gaussian mixture models (GMMs) learning, image similar patches can be clustered and the subspaces can be learned. The learned GMMs from clean images are then used to guide the clustering of noisypatches of the input noisy images, followed by a low-rank approximation process to estimate the latent subspace for image recovery. Numerical experiments show that the proposed method outperforms many state-of-the-art denoising algorithms such as BM3D and WNNM.
Similar papers:
  • Harvesting Discriminative Meta Objects With Deep CNN Features for Scene Classification [pdf] - Ruobing Wu, Baoyuan Wang, Wenping Wang, Yizhou Yu
  • Learning Parametric Distributions for Image Super-Resolution: Where Patch Matching Meets Sparse Coding [pdf] - Yongbo Li, Weisheng Dong, Guangming Shi, Xuemei Xie
  • Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf] - Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • Low-Rank Tensor Approximation With Laplacian Scale Mixture Modeling for Multiframe Image Denoising [pdf] - Weisheng Dong, Guangyu Li, Guangming Shi, Xin Li, Yi Ma
  • Patch Group Based Nonlocal Self-Similarity Prior Learning for Image Denoising [pdf] - Jun Xu, Lei Zhang, Wangmeng Zuo, David Zhang, Xiangchu Feng
Integrating Dashcam Views Through Inter-Video Mapping [pdf]
Hsin-I Chen, Yi-Ling Chen, Wei-Tse Lee, Fan Wang, Bing-Yu Chen

Abstract: In this paper, an inter-video mapping approach is proposed to integrate video footages from two dashcams installed on a preceding and its following vehicle to provide the illusion that the driver of the following vehicle can seethrough the preceding one. The key challenge is to adapt the perspectives of the two videos based on a small number of common features since a large portion of the common region in the video captured by the following vehicle is occluded by the preceding one. Inspired by the observation that images with the most similar viewpoints yield dense and high-quality matches, the proposed inter-video mapping estimates spatially-varying motions across the two videos utilizing images of very similar contents. Specifically, we estimate frame-to-frame motions of each two consecutive images and incrementally add new views to form long-range motion representation. On the other hand, we dynamically infer spatial-varying motions across the two videos that propagated from local feature correspondences to trajectories. In this way, the observed perspective discrepancy between the two videos can be well approximated by our motion estimation. Once the inter-video mapping is established, the correspondences can be updated incrementally, so the proposed method is suitable for on-line applications. Experiments with real-world challenging videos demonstrate the effectiveness of our approach.
Similar papers:
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
  • Sparse Dynamic 3D Reconstruction From Unsynchronized Videos [pdf] - Enliang Zheng, Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm
  • Robust Non-Rigid Motion Tracking and Surface Reconstruction Using L0 Regularization [pdf] - Kaiwen Guo, Feng Xu, Yangang Wang, Yebin Liu, Qionghai Dai
  • Motion Trajectory Segmentation via Minimum Cost Multicuts [pdf] - Margret Keuper, Bjoern Andres, Thomas Brox
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
Robust Nonrigid Registration by Convex Optimization [pdf]
Qifeng Chen, Vladlen Koltun

Abstract: We present an approach to nonrigid registration of 3D surfaces. We cast isometric embedding as MRF optimization and apply efficient global optimization algorithms based on linear programming relaxations. The Markov random field perspective suggests a natural connection with robust statistics and motivates robust forms of the intrinsic distortion functional. Our approach outperforms a large body of prior work by a significant margin, increasing registration precision on real data by a factor of 3.
Similar papers:
  • Registering Images to Untextured Geometry Using Average Shading Gradients [pdf] - Tobias Plotz, Stefan Roth
  • BodyPrint: Pose Invariant 3D Shape Matching of Human Bodies [pdf] - Jiangping Wang, Kai Ma, Vivek Kumar Singh, Thomas Huang, Terrence Chen
  • Detailed Full-Body Reconstructions of Moving People From Monocular RGB-D Sequences [pdf] - Federica Bogo, Michael J. Black, Matthew Loper, Javier Romero
  • A Groupwise Multilinear Correspondence Optimization for 3D Faces [pdf] - Timo Bolkart, Stefanie Wuhrer
  • An Adaptive Data Representation for Robust Point-Set Registration and Merging [pdf] - Dylan Campbell, Lars Petersson
Webly Supervised Learning of Convolutional Networks [pdf]
Xinlei Chen, Abhinav Gupta

Abstract: chihuahua jasmine We present an approach to utilize large amounts of web data for learning CNNs. Specifically inspired by curriculum learning, we present a two-step approach for CNN training. First, we use easy images to train an initial visual representation. We then use this initial CNN and adapt it to harder, more realistic images by leveraging the structure of data and categories. We demonstrate that our two-stage CNN outperforms a fine-tuned CNN trained on ImageNet on Pascal VOC 2012. We also demonstrate the strength of webly supervised learning by localizing objects in web images and training a R-CNN style [19] detector. It achieves the best performance on VOC 2007 where no VOC training data is used. Finally, we show our approach is quite robust to noise and performs comparably even when we use image search results from March 2013 (pre-CNN image search era).
Similar papers:
  • Towards Computational Baby Learning: A Weakly-Supervised Approach for Object Detection [pdf] - Xiaodan Liang, Si Liu, Yunchao Wei, Luoqi Liu, Liang Lin, Shuicheng Yan
  • Automatic Concept Discovery From Parallel Text and Visual Corpora [pdf] - Chen Sun, Chuang Gan, Ram Nevatia
  • Unsupervised Learning of Visual Representations Using Videos [pdf] - Xiaolong Wang, Abhinav Gupta
  • Augmenting Strong Supervision Using Web Data for Fine-Grained Categorization [pdf] - Zhe Xu, Shaoli Huang, Ya Zhang, Dacheng Tao
  • Relaxing From Vocabulary: Robust Weakly-Supervised Deep Learning for Vocabulary-Free Image Tagging [pdf] - Jianlong Fu, Yue Wu, Tao Mei, Jinqiao Wang, Hanqing Lu, Yong Rui
An Exploration of Parameter Redundancy in Deep Networks With Circulant Projections [pdf]
Yu Cheng, Felix X. Yu, Rogerio S. Feris, Sanjiv Kumar, Alok Choudhary, Shi-Fu Chang

Abstract: We explore the redundancy of parameters in deep neural networks by replacing the conventional linear projection in fully-connected layers with the circulant projection. The circulant structure substantially reduces memory footprint and enables the use of the Fast Fourier Transform to speed up the computation. Considering a fullyconnected neural network layer with d input nodes, and d output nodes, this method improves the time complexity from O(d2 ) to O(d log d) and space complexity from O(d2 ) to O(d). The space savings are particularly important for modern deep convolutional neural network architectures, where fully-connected layers typically contain more than 90% of the network parameters. We further show that the gradient computation and optimization of the circulant projections can be performed very efficiently. Our experiments on three standard datasets show that the proposed approach achieves this significant gain in storage and efficiency with minimal increase in error rate compared to neural networks with unstructured projections.
Similar papers:
  • Fast Orthogonal Projection Based on Kronecker Product [pdf] - Xu Zhang, Felix X. Yu, Ruiqi Guo, Sanjiv Kumar, Shengjin Wang, Shi-Fu Chang
  • Matrix Backpropagation for Deep Networks With Structured Layers [pdf] - Catalin Ionescu, Orestis Vantzos, Cristian Sminchisescu
  • Learning The Structure of Deep Convolutional Networks [pdf] - Jiashi Feng, Trevor Darrell
  • Fast R-CNN [pdf] - Ross Girshick
  • Deep Fried Convnets [pdf] - Zichao Yang, Marcin Moczulski, Misha Denil, Nando de Freitas, Alex Smola, Le Song, Ziyu Wang
Beyond White: Ground Truth Colors for Color Constancy Correction [pdf]
Dongliang Cheng, Brian Price, Scott Cohen, Michael S. Brown

Abstract: A limitation in color constancy research is the inability to establish ground truth colors for evaluating corrected images. Many existing datasets contain images of scenes with a color chart included; however, only the chart’s neutral colors (grayscale patches) are used to provide the ground truth for illumination estimation and correction. This is because the corrected neutral colors are known to lie along the achromatic line in the camera’s color space (i.e. R=G=B) ; the correct RGB values of the other color patches are not known. As a result, most methods estimate a 3×3 diagonal matrix that ensures only the neutral colors are correct. In this paper, we describe how to overcome this limitation. Specifically, we show that under certain illuminations, a diagonal 3×3 matrix is capable of correcting not only neutral colors, but all the colors in a scene. This finding allows us to find the ground truth RGB values for the color chart in the camera’s color space. We show how to use this information to correct all the images in existing datasets to have correct colors. Working from these new color corrected datasets, we describe how to modify existing color constancy algorithms to perform better image correction.
Similar papers:
  • Synthesizing Illumination Mosaics From Internet Photo-Collections [pdf] - Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm
  • Illumination Robust Color Naming via Label Propagation [pdf] - Yuanliu liu, Zejian Yuan, Badong Chen, Jianru Xue, Nanning Zheng
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
  • Convolutional Color Constancy [pdf] - Jonathan T. Barron
Deep Colorization [pdf]
Zezhou Cheng, Qingxiong Yang, Bin Sheng

Abstract: This paper investigates into the colorization problem which converts a grayscale image to a colorful version. This is a very difficult problem and normally requires manual adjustment to achieve artifact-free quality. For instance, it normally requires human-labelled color scribbles on the grayscale target image or a careful selection of colorful reference images (e.g., capturing the same scene in the grayscale target image). Unlike the previous methods, this paper aims at a high-quality fully-automatic colorization method. With the assumption of a perfect patch matching technique, the use of an extremely large-scale reference database (that contains sufficient color images) is the most reliable solution to the colorization problem. However, patch matching noise will increase with respect to the size of the reference database in practice. Inspired by the recent success in deep learning techniques which provide amazing modeling of large-scale data, this paper reformulates the colorization problem so that deep learning techniques can be directly employed. To ensure artifactfree quality, a joint bilateral filtering based post-processing step is proposed. Numerous experiments demonstrate that our method outperforms the state-of-art algorithms both in terms of quality and speed.
Similar papers:
  • SOWP: Spatially Ordered and Weighted Patch Descriptor for Visual Tracking [pdf] - Han-Ul Kim, Dae-Youn Lee, Jae-Young Sim, Chang-Su Kim
  • Visual Tracking With Fully Convolutional Networks [pdf] - Lijun Wang, Wanli Ouyang, Xiaogang Wang, Huchuan Lu
  • Predicting Good Features for Image Geo-Localization Using Per-Bundle VLAD [pdf] - Hyo Jin Kim, Enrique Dunn, Jan-Michael Frahm
  • Synthesizing Illumination Mosaics From Internet Photo-Collections [pdf] - Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm
  • Mutual-Structure for Joint Filtering [pdf] - Xiaoyong Shen, Chao Zhou, Li Xu, Jiaya Jia
Query Adaptive Similarity Measure for RGB-D Object Recognition [pdf]
Yanhua Cheng, Rui Cai, Chi Zhang, Zhiwei Li, Xin Zhao, Kaiqi Huang, Yong Rui

Abstract: This paper studies the problem of improving the top-1 accuracy of RGB-D object recognition. Despite of the impressive top-5 accuracies achieved by existing methods, their top-1 accuracies are not very satisfactory. The reasons are in two-fold: (1) existing similarity measures are sensitive to object pose and scale changes, as well as intra-class variations; and (2) effectively fusing RGB and depth cues is still an open problem. To address these problems, this paper first proposes a new similarity measure based on dense matching, through which objects in comparison are warped and aligned, to better tolerate variations. Towards RGB and depth fusion, we argue that a constant and golden weight doesnt exist. The two modalities have varying contributions when comparing objects from different categories. To capture such a dynamic characteristic, a group of matchers equipped with various fusion weights is constructed, to explore the responses of dense matching under different fusion configurations. All the response scores are finally merged following a learning-to-combination way, which provides quite good generalization ability in practice. The proposed approach win the best results on several public benchmarks, e.g., achieves 92.7% top-1 test accuracy on the Washington RGB-D object dataset, with a 5.1% improvement over the state-of-the-art.
Similar papers:
  • MMSS: Multi-Modal Sharable and Specific Feature Learning for RGB-D Object Recognition [pdf] - Anran Wang, Jianfei Cai, Jiwen Lu, Tat-Jen Cham
  • High Quality Structure From Small Motion for Rolling Shutter Cameras [pdf] - Sunghoon Im, Hyowon Ha, Gyeongmin Choe, Hae-Gon Jeon, Kyungdon Joo, In So Kweon
  • Predicting Good Features for Image Geo-Localization Using Per-Bundle VLAD [pdf] - Hyo Jin Kim, Enrique Dunn, Jan-Michael Frahm
  • Learning Query and Image Similarities With Ranking Canonical Correlation Analysis [pdf] - Ting Yao, Tao Mei, Chong-Wah Ngo
  • Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf] - Naejin Kong, Michael J. Black
P-CNN: Pose-Based CNN Features for Action Recognition [pdf]
Guilhem Cheron, Ivan Laptev, Cordelia Schmid

Abstract: This work targets human action recognition in video. While recent methods typically represent actions by statistics of local video features, here we argue for the importance of a representation derived from human pose. To this end we propose a new Pose-based Convolutional Neural Network descriptor (P-CNN) for action recognition. The descriptor aggregates motion and appearance information along tracks of human body parts. We investigate different schemes of temporal aggregation and experiment with P-CNN features obtained both for automatically estimated and manually annotated human poses. We evaluate our method on the recent and challenging JHMDB and MPII Cooking datasets. For both datasets our method shows consistent improvement over the state of the art.
Similar papers:
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
  • Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf] - Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • Flowing ConvNets for Human Pose Estimation in Videos [pdf] - Tomas Pfister, James Charles, Andrew Zisserman
Semi-Supervised Normalized Cuts for Image Segmentation [pdf]
Selene E. Chew, Nathan D. Cahill

Abstract: Since its introduction as a powerful graph-based method for image segmentation, the Normalized Cuts (NCuts) algorithm has been generalized to incorporate expert knowledge about how certain pixels or regions should be grouped, or how the resulting segmentation should be biased to be correlated with priors. Previous approaches incorporate hard must-link constraints on how certain pixels should be grouped as well as hard cannot-link constraints on how other pixels should be separated into different groups. In this paper, we reformulate NCuts to allow both sets of constraints to be handled in a soft manner, enabling the user to tune the degree to which the constraints are satised. An approximate spectral solution to the reformulated problem exists without requiring explicit construction of a large, dense matrix; hence, computation time is comparable to that of unconstrained NCuts. Using synthetic data and real imagery, we show that soft handling of constraints yields better results than unconstrained NCuts and enables more robust clustering and segmentation than is possible when the constraints are strictly enforced.
Similar papers:
  • Efficient Video Segmentation Using Parametric Graph Partitioning [pdf] - Chen-Ping Yu, Hieu Le, Gregory Zelinsky, Dimitris Samaras
  • Semantic Component Analysis [pdf] - Calvin Murdock, Fernando De la Torre
  • A Projection Free Method for Generalized Eigenvalue Problem With a Nonsmooth Regularizer [pdf] - Seong Jae Hwang, Maxwell D. Collins, Sathya N. Ravi, Vamsi K. Ithapu, Nagesh Adluru, Sterling C. Johnson, Vikas Singh
  • Semantic Segmentation of RGBD Images With Mutex Constraints [pdf] - Zhuo Deng, Sinisa Todorovic, Longin Jan Latecki
  • Constrained Convolutional Neural Networks for Weakly Supervised Segmentation [pdf] - Deepak Pathak, Philipp Krahenbuhl, Trevor Darrell
See the Difference: Direct Pre-Image Reconstruction and Pose Estimation by Differentiating HOG [pdf]
Wei-Chen Chiu, Mario Fritz

Abstract: pre-image reconstruction The Histogram of Oriented Gradient (HOG) descriptor has led to many advances in computer vision over the last decade and is still part of many state of the art approaches. We realize that the associated feature computation is piecewise differentiable and therefore many pipelines which build on HOG can be made differentiable. This lends to advanced introspection as well as opportunities for end-to-end optimization. We present our implementation of HOG based on the auto-differentiation toolbox Chumpy [18] and show applications to pre-image visualization and pose estimation which extends the existing differentiable renderer OpenDR [19] pipeline. Both applications improve on the respective state-of-the-art HOG approaches. ! ! ! ! E = !(I) (I) ! (I) HOG (I) diff.HOG pose estimation @E @ I I Pf E @E @f diff.HOG HOG discriminative patches PI observation I locations Pf rendered image I f @f @U 2D projected vertex coordinates camera parameters I = PI Pf similarity E PI U C distance to camera V A @U @V @V @ azimuth elevation
Similar papers:
  • Pose Induction for Novel Object Categories [pdf] - Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • Real-Time Pose Estimation Piggybacked on Object Detection [pdf] - Roman Juranek, Adam Herout, Marketa Dubska, Pavel Zemcik
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • Registering Images to Untextured Geometry Using Average Shading Gradients [pdf] - Tobias Plotz, Stefan Roth
  • Learning Analysis-by-Synthesis for 6D Pose Estimation in RGB-D Images [pdf] - Alexander Krull, Eric Brachmann, Frank Michel, Michael Ying Yang, Stefan Gumhold, Carsten Rother
A Collaborative Filtering Approach to Real-Time Hand Pose Estimation [pdf]
Chiho Choi, Ayan Sinha, Joon Hee Choi, Sujin Jang, Karthik Ramani

Abstract: Collaborative filtering aims to predict unknown user ratings in a recommender system by collectively assessing known user preferences. In this paper, we first draw analogies between collaborative filtering and the pose estimation problem. Specifically, we recast the hand pose estimation problem as the cold-start problem for a new user with unknown item ratings in a recommender system. Inspired by fast and accurate matrix factorization techniques for collaborative filtering, we develop a real-time algorithm for estimating the hand pose from RGB-D data of a commercial depth camera. First, we efficiently identify nearest neighbors using local shape descriptors in the RGB-D domain from a library of hand poses with known pose parameter values. We then use this information to evaluate the unknown pose parameters using a joint matrix factorization and completion (JMFC) approach. Our quantitative and qualitative results suggest that our approach is robust to variation in hand configurations while achieving real time performance ( 29 FPS) on a standard computer.
Similar papers:
  • 3D Hand Pose Estimation Using Randomized Decision Forest With Segmentation Index Points [pdf] - Peiyi Li, Haibin Ling, Xi Li, Chunyuan Liao
  • Opening the Black Box: Hierarchical Sampling Optimization for Estimating Human Hand Pose [pdf] - Danhang Tang, Jonathan Taylor, Pushmeet Kohli, Cem Keskin, Tae-Kyun Kim, Jamie Shotton
  • Understanding Everyday Hands in Action From RGB-D Images [pdf] - Gregory Rogez, James S. Supancic III, Deva Ramanan
  • Depth-Based Hand Pose Estimation: Data, Methods, and Challenges [pdf] - James S. Supancic III, Gregory Rogez, Yi Yang, Jamie Shotton, Deva Ramanan
  • Training a Feedback Loop for Hand Pose Estimation [pdf] - Markus Oberweger, Paul Wohlhart, Vincent Lepetit
Near-Online Multi-Target Tracking With Aggregated Local Flow Descriptor [pdf]
Wongun Choi

Abstract: In this paper, we tackle two key aspects of multiple target tracking problem: 1) designing an accurate affinity measure to associate detections and 2) implementing an efficient and accurate (near) online multiple target tracking algorithm. As for the first contribution, we introduce a novel Aggregated Local Flow Descriptor (ALFD) that encodes the relative motion pattern between a pair of temporally distant detections using long term interest point trajectories (IPTs). Leveraging on the IPTs, the ALFD provides a robust affinity measure for estimating the likelihood of matching detections regardless of the application scenarios. As for another contribution, we present a Near-Online Multi-target Tracking (NOMT) algorithm. The tracking problem is formulated as a data-association between targets and detections in a temporal window, that is performed repeatedly at every frame. While being efficient, NOMT achieves robustness via integrating multiple cues including ALFD metric, target dynamics, appearance similarity, and long term trajectory regularization into the model. Our ablative analysis verifies the superiority of the ALFD metric over the other conventional affinity metrics. We run a comprehensive experimental evaluation on two challenging tracking datasets, KITTI [16] and MOT [2] datasets. The NOMT method combined with ALFD metric achieves the best accuracy in both datasets with significant margins (about 10% higher MOTA) over the state-of-the-art.
Similar papers:
  • Multiple Hypothesis Tracking Revisited [pdf] - Chanho Kim, Fuxin Li, Arridhana Ciptadi, James M. Rehg
  • FollowMe: Efficient Online Min-Cost Flow Tracking With Bounded Memory and Computation [pdf] - Philip Lenz, Andreas Geiger, Raquel Urtasun
  • Joint Probabilistic Data Association Revisited [pdf] - Seyed Hamid Rezatofighi, Anton Milan, Zhen Zhang, Qinfeng Shi, Anthony Dick, Ian Reid
  • Learning to Divide and Conquer for Online Multi-Target Tracking [pdf] - Francesco Solera, Simone Calderara, Rita Cucchiara
  • Learning to Track: Online Multi-Object Tracking by Decision Making [pdf] - Yu Xiang, Alexandre Alahi, Silvio Savarese
Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf]
Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang

Abstract: In this paper, we propose to predict immediacy for interacting persons from still images. A complete immediacy set includes interactions, relative distance, body leaning direction and standing orientation. These measures are found to be related to the attitude, social relationship, social interaction, action, nationality, and religion of the communicators. 1 A large-scale dataset with 10, 000 images is constructed, in which all the immediacy cues and the human poses are annotated. We propose a rich set of immediacy representations that help to predict immediacy from imperfect 1-person and 2-person pose estimation results. A multi-task deep recurrent neural network is constructed to take the proposed rich immediacy representations as the input and learn the complex relationship among immediacy predictions through multiple steps of refinement. The effectiveness of the proposed approach is proved through extensive experiments on the large-scale dataset.
Similar papers:
  • Recurrent Network Models for Human Dynamics [pdf] - Katerina Fragkiadaki, Sergey Levine, Panna Felsen, Jitendra Malik
  • Human Pose Estimation in Videos [pdf] - Dong Zhang, Mubarak Shah
  • Pose Induction for Novel Object Categories [pdf] - Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • Uncovering Interactions and Interactors: Joint Estimation of Head, Body Orientation and F-Formations From Surveillance Videos [pdf] - Elisa Ricci, Jagannadan Varadarajan, Ramanathan Subramanian, Samuel Rota Bulo, Narendra Ahuja, Oswald Lanz
Unsupervised Synchrony Discovery in Human Interaction [pdf]
Wen-Sheng Chu, Jiabei Zeng, Fernando De la Torre, Jeffrey F. Cohn, Daniel S. Messinger

Abstract: People are inherently social. Social interaction plays an important and natural role in human behavior. Most computational methods focus on individuals alone rather than in social context. They also require labelled training data. We present an unsupervised approach to discover interpersonal synchrony, referred as to two or more persons preforming common actions in overlapping video frames or segments. For computational efficiency, we develop a branch-and-bound (B&B) approach that affords exhaustive search while guaranteeing a globally optimal solution. The proposed method is entirely general. It takes from two or more videos any multi-dimensional signal that can be represented as a histogram. We derive three novel bounding functions and provide efficient extensions, including multisynchrony detection and accelerated search, using a warmstart strategy and parallelism. We evaluate the effectiveness of our approach in multiple databases, including human actions using the CMU Mocap dataset [1], spontaneous facial behaviors using group-formation task dataset [37] and parent-infant interaction dataset [28].
Similar papers:
  • Unsupervised Semantic Parsing of Video Collections [pdf] - Ozan Sener, Amir R. Zamir, Silvio Savarese, Ashutosh Saxena
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Unsupervised Object Discovery and Tracking in Video Collections [pdf] - Suha Kwak, Minsu Cho, Ivan Laptev, Jean Ponce, Cordelia Schmid
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
You Are Here: Mimicking the Human Thinking Process in Reading Floor-Plans [pdf]
Hang Chu, Dong Ki Kim, Tsuhan Chen

Abstract: A human can easily find his or her way in an unfamiliar building, by walking around and reading the floor-plan. We try to mimic and automate this human thinking process. More precisely, we introduce a new and useful task of locating an user in the floor-plan, by using only a camera and a floor-plan without any other prior information. We address the problem with a novel matching-localization algorithm that is inspired by human logic. We demonstrate through experiments that our method outperforms state-of-the-art floor-plan-based localization methods by a large margin, while also being highly efficient for real-time applications.
Similar papers:
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
  • Camera Pose Voting for Large-Scale Image-Based Localization [pdf] - Bernhard Zeisl, Torsten Sattler, Marc Pollefeys
  • PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization [pdf] - Alex Kendall, Matthew Grimes, Roberto Cipolla
  • Lost Shopping! Monocular Localization in Large Indoor Spaces [pdf] - Shenlong Wang, Sanja Fidler, Raquel Urtasun
  • Localize Me Anywhere, Anytime: A Multi-Task Point-Retrieval Approach [pdf] - Guoyu Lu, Yan Yan, Li Ren, Jingkuan Song, Nicu Sebe, Chandra Kambhamettu
Low Dimensional Explicit Feature Maps [pdf]
Ondrej Chum

Abstract: Approximating non-linear kernels1 by finite-dimensional feature maps is a popular approach for speeding up training and evaluation of support vector machines or to encode information into efficient match kernels. We propose a novel method of data independent construction of low dimensional feature maps. The problem is cast as a linear program which jointly considers competing objectives: the quality of the approximation and the dimensionality of the feature map. For both shift-invariant and homogeneous kernels the proposed method achieves a better approximations at the same dimensionality or comparable approximations at lower dimensionality of the feature map compared with state-of-the-art methods.
Similar papers:
  • Class-Specific Image Deblurring [pdf] - Saeed Anwar, Cong Phuoc Huynh, Fatih Porikli
  • Beyond Covariance: Feature Representation With Nonlinear Kernel Matrices [pdf] - Lei Wang, Jianjia Zhang, Luping Zhou, Chang Tang, Wanqing Li
  • Secrets of GrabCut and Kernel K-Means [pdf] - Meng Tang, Ismail Ben Ayed, Dmitrii Marin, Yuri Boykov
  • Per-Sample Kernel Adaptation for Visual Recognition and Grouping [pdf] - Borislav Antic, Bjorn Ommer
  • Additive Nearest Neighbor Feature Maps [pdf] - Zhenzhen Wang, Xiao-Tong Yuan, Qingshan Liu, Shuicheng Yan
Minimizing Human Effort in Interactive Tracking by Incremental Learning of Model Parameters [pdf]
Arridhana Ciptadi, James M. Rehg

Abstract: We address the problem of minimizing human effort in interactive tracking by learning sequence-specific model parameters. Determining the optimal model parameters for each sequence is a critical problem in tracking. We demonstrate that by using the optimal model parameters for each sequence we can achieve high precision tracking results with significantly less effort. We leverage the sequential nature of interactive tracking to formulate an efficient method for learning model parameters through a maximum margin framework. By using our method we are able to save 60 90% of human effort to achieve high precision on two datasets: the VIRAT dataset and an Infant-Mother Interaction dataset.
Similar papers:
  • Model-Based Tracking at 300Hz Using Raw Time-of-Flight Observations [pdf] - Jan Stuhmer, Sebastian Nowozin, Andrew Fitzgibbon, Richard Szeliski, Travis Perry, Sunil Acharya, Daniel Cremers, Jamie Shotton
  • FollowMe: Efficient Online Min-Cost Flow Tracking With Bounded Memory and Computation [pdf] - Philip Lenz, Andreas Geiger, Raquel Urtasun
  • A Versatile Learning-Based 3D Temporal Tracker: Scalable, Robust, Online [pdf] - David Joseph Tan, Federico Tombari, Slobodan Ilic, Nassir Navab
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
  • Joint Probabilistic Data Association Revisited [pdf] - Seyed Hamid Rezatofighi, Anton Milan, Zhen Zhang, Qinfeng Shi, Anthony Dick, Ian Reid
Merging the Unmatchable: Stitching Visually Disconnected SfM Models [pdf]
Andrea Cohen, Torsten Sattler, Marc Pollefeys

Abstract: Recent advances in Structure-from-Motion not only enable the reconstruction of large scale scenes, but are also able to detect ambiguous structures caused by repeating elements that might result in incorrect reconstructions. Yet, it is not always possible to fully reconstruct a scene. The images required to merge different sub-models might be missing or it might be impossible to acquire such images in the first place due to occlusions or the structure of the scene. The problem of aligning multiple reconstructions that do not have visual overlap is impossible to solve in general. An important variant of this problem is the case in which individual sides of a building can be reconstructed but not joined due to the missing visual overlap. In this paper, we present a combinatorial approach for solving this variant by automatically stitching multiple sides of a building together. Our approach exploits symmetries and semantic information to reason about the possible geometric relations between the individual models. We show that our approach is able to reconstruct complete building models where traditional SfM ends up with disconnected building sides.
Similar papers:
  • Unsupervised Generation of a Viewpoint Annotated Car Dataset From Videos [pdf] - Nima Sedaghat, Thomas Brox
  • Global Structure-From-Motion by Similarity Averaging [pdf] - Zhaopeng Cui, Ping Tan
  • Exploiting Object Similarity in 3D Reconstruction [pdf] - Chen Zhou, Fatma Guney, Yizhou Wang, Andreas Geiger
  • Detection and Segmentation of 2D Curved Reflection Symmetric Structures [pdf] - Ching L. Teo, Cornelia Fermuller, Yiannis Aloimonos
  • Procedural Editing of 3D Building Point Clouds [pdf] - Ilke Demir, Daniel G. Aliaga, Bedrich Benes
The Likelihood-Ratio Test and Efficient Robust Estimation [pdf]
Andrea Cohen, Christopher Zach

Abstract: Robust estimation of model parameters in the presence of outliers is a key problem in computer vision. RANSAC inspired techniques are widely used in this context, although their application might be limited due to the need of a priori knowledge on the inlier noise level. We propose a new approach for jointly optimizing over model parameters and the inlier noise level based on the likelihood ratio test. This allows control over the type I error incurred. We also propose an early bailout strategy for efficiency. Tests on both synthetic and real data show that our method outperforms the state-of-the-art in a fraction of the time.
Similar papers:
  • Globally Optimal 2D-3D Registration From Points or Lines Without Correspondences [pdf] - Mark Brown, David Windridge, Jean-Yves Guillemaut
  • Hyperpoints and Fine Vocabularies for Large-Scale Location Recognition [pdf] - Torsten Sattler, Michal Havlena, Filip Radenovic, Konrad Schindler, Marc Pollefeys
  • Robust and Optimal Sum-of-Squares-Based Point-to-Plane Registration of Image Sets and Structured Scenes [pdf] - Danda Pani Paudel, Adlane Habed, Cedric Demonceaux, Pascal Vasseur
  • Low-Rank Matrix Factorization Under General Mixture Noise Distributions [pdf] - Xiangyong Cao, Yang Chen, Qian Zhao, Deyu Meng, Yao Wang, Dong Wang, Zongben Xu
  • An Efficient Statistical Method for Image Noise Level Estimation [pdf] - Guangyong Chen, Fengyuan Zhu, Pheng Ann Heng
A Novel Representation of Parts for Accurate 3D Object Detection and Tracking in Monocular Images [pdf]
Alberto Crivellaro, Mahdi Rad, Yannick Verdie, Kwang Moo Yi, Pascal Fua, Vincent Lepetit

Abstract: We present a method that estimates in real-time and under challenging conditions the 3D pose of a known object. Our method relies only on grayscale images since depth cameras fail on metallic objects; it can handle poorly textured objects, and cluttered, changing environments; the pose it predicts degrades gracefully in presence of large occlusions. As a result, by contrast with the state-of-the-art, our method is suitable for practical Augmented Reality applications even in industrial environments. To be robust to occlusions, we first learn to detect some parts of the target object. Our key idea is to then predict the 3D pose of each part in the form of the 2D projections of a few control points. The advantages of this representation is three-fold: We can predict the 3D pose of the object even when only one part is visible; when several parts are visible, we can combine them easily to compute a better pose of the object; the 3D pose we obtain is usually very accurate, even when only few parts are visible.
Similar papers:
  • Training a Feedback Loop for Hand Pose Estimation [pdf] - Markus Oberweger, Paul Wohlhart, Vincent Lepetit
  • Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf] - Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang
  • Pose Induction for Novel Object Categories [pdf] - Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • Learning Where to Position Parts in 3D [pdf] - Marco Pedersoli, Tinne Tuytelaars
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
Global Structure-From-Motion by Similarity Averaging [pdf]
Zhaopeng Cui, Ping Tan

Abstract: Global structure-from-motion (SfM) methods solve all cameras simultaneously from all available relative motions. It has better potential in both reconstruction accuracy and computation efficiency than incremental methods. However, global SfM is challenging, mainly because of two reasons. Firstly, translation averaging is difficult, since an essential matrix only tells the direction of relative translation. Secondly, it is also hard to filter out bad essential matrices due to feature matching failures. We propose to compute a sparse depth image at each camera to solve both problems. Depth images help to upgrade an essential matrix to a similarity transformation, which can determine the scale of relative translation. Thus, camera registration is formulated as a well-posed similarity averaging problem. Depth images also make the filtering of essential matrices simple and effective. In this way, translation averaging can be solved robustly in two convex L1 optimization problems, which reach the global optimum rapidly. We demonstrate this method in various examples including sequential data, Internet data, and ambiguous data with repetitive scene structures.
Similar papers:
  • SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf] - Kensei Jo, Mohit Gupta, Shree K. Nayar
  • Structure From Motion Using Structure-Less Resection [pdf] - Enliang Zheng, Changchang Wu
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
  • An Efficient Minimal Solution for Multi-Camera Motion [pdf] - Jonathan Ventura, Clemens Arth, Vincent Lepetit
  • High Quality Structure From Small Motion for Rolling Shutter Cameras [pdf] - Sunghoon Im, Hyowon Ha, Gyeongmin Choe, Hae-Gon Jeon, Kyungdon Joo, In So Kweon
BoxSup: Exploiting Bounding Boxes to Supervise Convolutional Networks for Semantic Segmentation [pdf]
Jifeng Dai, Kaiming He, Jian Sun

Abstract: Recent leading approaches to semantic segmentation rely on deep convolutional networks trained with humanannotated, pixel-level segmentation masks. Such pixelaccurate supervision demands expensive labeling effort and limits the performance of deep networks that usually benefit from more training data. In this paper, we propose a method that achieves competitive accuracy but only requires easily obtained bounding box annotations. The basic idea is to iterate between automatically generating region proposals and training convolutional networks. These two steps gradually recover segmentation masks for improving the networks, and vise versa. Our method, called BoxSup, produces competitive results (e.g., 62.0% mAP for validation) supervised by boxes only, on par with strong baselines (e.g., 63.8% mAP) fully supervised by masks under the same setting. By leveraging a large amount of bounding boxes, BoxSup further yields state-of-the-art results on PASCAL VOC 2012 and PASCAL-CONTEXT [26].
Similar papers:
  • Joint Object and Part Segmentation Using Deep Learned Potentials [pdf] - Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille
  • Learning Deconvolution Network for Semantic Segmentation [pdf] - Hyeonwoo Noh, Seunghoon Hong, Bohyung Han
  • Object Detection via a Multi-Region and Semantic Segmentation-Aware CNN Model [pdf] - Spyros Gidaris, Nikos Komodakis
  • Segment-Phrase Table for Semantic Segmentation, Visual Entailment and Paraphrasing [pdf] - Hamid Izadinia, Fereshteh Sadeghi, Santosh K. Divvala, Hannaneh Hajishirzi, Yejin Choi, Ali Farhadi
  • Weakly- and Semi-Supervised Learning of a Deep Convolutional Network for Semantic Image Segmentation [pdf] - George Papandreou, Liang-Chieh Chen, Kevin P. Murphy, Alan L. Yuille
Fully Connected Guided Image Filtering [pdf]
Longquan Dai, Mengke Yuan, Feihu Zhang, Xiaopeng Zhang

Abstract: This paper presents a linear time fully connected guided filter by introducing the minimum spanning tree (MST) to the guided filter (GF). Since the intensity based filtering kernel of GF is apt to overly smooth edges and the fixed-shape local box support region adopted by GF is not geometricadaptive, our filter introduces an extra spatial term, the tree similarity, to the filtering kernel of GF and substitutes the box window with the implicit support region by establishing all-pairs-connections among pixels in the image and assigning the spatial-intensity-aware similarity to these connections. The adaptive implicit support region composed by the pixels with large kernel weights in the entire image domain has a big advantage over the predefined local box window in presenting the structure of an image for the reason that: 1, MST can efficiently present the structure of an image; 2, the kernel weight of our filter considers the tree distance defined on the MST. Due to these reasons, our filter achieves better edge-preserving results. We demonstrate the strength of the proposed filter in several applications. Experimental results show that our method produces better results than state-of-the-art methods.
Similar papers:
  • Semantic Segmentation of RGBD Images With Mutex Constraints [pdf] - Zhuo Deng, Sinisa Todorovic, Longin Jan Latecki
  • Object Detection via a Multi-Region and Semantic Segmentation-Aware CNN Model [pdf] - Spyros Gidaris, Nikos Komodakis
  • Adaptive Exponential Smoothing for Online Filtering of Pixel Prediction Maps [pdf] - Kang Dang, Jiong Yang, Junsong Yuan
  • Mutual-Structure for Joint Filtering [pdf] - Xiaoyong Shen, Chao Zhou, Li Xu, Jiaya Jia
  • Segment Graph Based Image Filtering: Fast Structure-Preserving Smoothing [pdf] - Feihu Zhang, Longquan Dai, Shiming Xiang, Xiaopeng Zhang
Learning Spatially Regularized Correlation Filters for Visual Tracking [pdf]
Martin Danelljan, Gustav Hager, Fahad Shahbaz Khan, Michael Felsberg

Abstract: Robust and accurate visual tracking is one of the most challenging computer vision problems. Due to the inherent lack of training data, a robust approach for constructing a target appearance model is crucial. Recently, discriminatively learned correlation filters (DCF) have been successfully applied to address this problem for tracking. These methods utilize a periodic assumption of the training samples to efficiently learn a classifier on all patches in the target neighborhood. However, the periodic assumption also introduces unwanted boundary effects, which severely degrade the quality of the tracking model. We propose Spatially Regularized Discriminative Correlation Filters (SRDCF) for tracking. A spatial regularization component is introduced in the learning to penalize correlation filter coefficients depending on their spatial location. Our SRDCF formulation allows the correlation filters to be learned on a significantly larger set of negative training samples, without corrupting the positive samples. We further propose an optimization strategy, based on the iterative Gauss-Seidel method, for efficient online learning of our SRDCF. Experiments are performed on four benchmark datasets: OTB-2013, ALOV++, OTB-2015, and VOT2014. Our approach achieves state-of-the-art results on all four datasets. On OTB-2013 and OTB-2015, we obtain an absolute gain of 8.0% and 8.2% respectively, in mean overlap precision, compared to the best existing trackers.
Similar papers:
  • Online Object Tracking With Proposal Selection [pdf] - Yang Hua, Karteek Alahari, Cordelia Schmid
  • SOWP: Spatially Ordered and Weighted Patch Descriptor for Visual Tracking [pdf] - Han-Ul Kim, Dae-Youn Lee, Jae-Young Sim, Chang-Su Kim
  • Understanding and Diagnosing Visual Tracking Systems [pdf] - Naiyan Wang, Jianping Shi, Dit-Yan Yeung, Jiaya Jia
  • Multi-Kernel Correlation Filter for Visual Tracking [pdf] - Ming Tang, Jiayi Feng
  • Hierarchical Convolutional Features for Visual Tracking [pdf] - Chao Ma, Jia-Bin Huang, Xiaokang Yang, Ming-Hsuan Yang
Adaptive Exponential Smoothing for Online Filtering of Pixel Prediction Maps [pdf]
Kang Dang, Jiong Yang, Junsong Yuan

Abstract: We propose an efficient online video filtering method, called adaptive exponential filtering (AES) to refine pixel prediction maps. Assuming each pixel is associated with a discriminative prediction score, the proposed AES applies exponentially decreasing weights over time to smooth the prediction score of each pixel, similar to classic exponential smoothing. However, instead of fixing the spatial pixel location to perform temporal filtering, we trace each pixel in the past frames by finding the optimal path that can bring the maximum exponential smoothing score, thus performing adaptive and non-linear filtering. Thanks to the pixel tracing, AES can better address object movements and avoid over-smoothing. To enable real-time filtering, we propose a linear-complexity dynamic programming scheme that can trace all pixels simultaneously. We apply the proposed filtering method to improve both saliency detection maps and scene parsing maps. The comparisons with average and exponential filtering, as well as state-ofthe-art methods, validate that our AES can effectively refine the pixel prediction maps, without using the original video again.
Similar papers:
  • Fully Connected Guided Image Filtering [pdf] - Longquan Dai, Mengke Yuan, Feihu Zhang, Xiaopeng Zhang
  • Video Restoration Against Yin-Yang Phasing [pdf] - Xiaolin Wu, Zhenhao Li, Xiaowei Deng
  • SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf] - Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao
  • Human Parsing With Contextualized Convolutional Neural Network [pdf] - Xiaodan Liang, Chunyan Xu, Xiaohui Shen, Jianchao Yang, Si Liu, Jinhui Tang, Liang Lin, Shuicheng Yan
  • Segment Graph Based Image Filtering: Fast Structure-Preserving Smoothing [pdf] - Feihu Zhang, Longquan Dai, Shiming Xiang, Xiaopeng Zhang
Pairwise Conditional Random Forests for Facial Expression Recognition [pdf]
Arnaud Dapogny, Kevin Bailly, Severine Dubuisson

Abstract: Facial expression can be seen as the dynamic variation of ones appearance over time. Successful recognition thus involves finding representations of high-dimensional spatiotemporal patterns that can be generalized to unseen facial morphologies and variations of the expression dynamics. In this paper, we propose to learn Random Forests from heterogeneous derivative features (e.g. facial fiducial point movements or texture variations) upon pairs of images. Those forests are conditioned on the expression label of the first frame to reduce the variability of the ongoing expression transitions. When testing on a specific frame of a video, pairs are created between this current frame and the previous ones. Predictions for each previous frame are used to draw trees from Pairwise Conditional Random Forests (PCRF) whose pairwise outputs are averaged over time to produce robust estimates. As such, PCRF appears as a natural extension of Random Forests to learn spatio-temporal patterns, that leads to significant improvements over standard Random Forests as well as state-of-the-art approaches on several facial expression benchmarks. Introduction In the last decades, automatic facial expression recognition (FER) has attracted an increasing attention, as it is a fundamental step of many applications such as humancomputer interaction, or assistive healthcare technologies. A good survey covering FER can be found in [23]. There exists many impediments to successful deciphering of facial expr
Similar papers:
  • Learning to Transfer: Transferring Latent Task Structures and Its Application to Person-Specific Facial Action Unit Detection [pdf] - Timur Almaev, Brais Martinez, Michel Valstar
  • Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf] - Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic
  • FaceDirector: Continuous Control of Facial Performance in Video [pdf] - Charles Malleson, Jean-Charles Bazin, Oliver Wang, Derek Bradley, Thabo Beeler, Adrian Hilton, Alexander Sorkine-Hornung
  • From Emotions to Action Units With Hidden and Semi-Hidden-Task Learning [pdf] - Adria Ruiz, Joost Van de Weijer, Xavier Binefa
  • Joint Fine-Tuning in Deep Neural Networks for Facial Expression Recognition [pdf] - Heechul Jung, Sihaeng Lee, Junho Yim, Sunjeong Park, Junmo Kim
Procedural Editing of 3D Building Point Clouds [pdf]
Ilke Demir, Daniel G. Aliaga, Bedrich Benes

Abstract: Thanks to the recent advances in computational photography and remote sensing, point clouds of buildings are becoming increasingly available, yet their processing poses various challenges. In our work, we tackle the problem of point cloud completion and editing and we approach it via inverse procedural modeling. Contrary to the previous work, our approach operates directly on the point cloud without an intermediate triangulation. Our approach consists of 1) semi-automatic segmentation of the input point cloud with segment comparison and template matching to detect repeating structures, 2) a consensus-based voting schema and a pattern extraction algorithm to discover completed terminal geometry and their patterns of usage, all encoded into a context-free grammar, and 3) an interactive editing tool where the user can create new point clouds by using procedural copy and paste operations, and smart resizing. We demonstrate our approach on editing of building models with up to 1.8M points. In our implementation, preprocessing takes up to several minutes and a single editing operation needs from one second to one minute depending on the model size and the operation type.
Similar papers:
  • Joint Camera Clustering and Surface Segmentation for Large-Scale Multi-View Stereo [pdf] - Runze Zhang, Shiwei Li, Tian Fang, Siyu Zhu, Long Quan
  • Exploiting Object Similarity in 3D Reconstruction [pdf] - Chen Zhou, Fatma Guney, Yizhou Wang, Andreas Geiger
  • Global, Dense Multiscale Reconstruction for a Billion Points [pdf] - Benjamin Ummenhofer, Thomas Brox
  • Higher-Order CRF Structural Segmentation of 3D Reconstructed Surfaces [pdf] - Jingbo Liu, Jinglu Wang, Tian Fang, Chiew-Lan Tai, Long Quan
  • Merging the Unmatchable: Stitching Visually Disconnected SfM Models [pdf] - Andrea Cohen, Torsten Sattler, Marc Pollefeys
Semantic Segmentation of RGBD Images With Mutex Constraints [pdf]
Zhuo Deng, Sinisa Todorovic, Longin Jan Latecki

Abstract: In this paper, we address the problem of semantic scene segmentation of RGB-D images of indoor scenes. We propose a novel image region labeling method which augments CRF formulation with hard mutual exclusion (mutex) constraints. This way our approach can make use of rich and accurate 3D geometric structure coming from Kinect in a principled manner. The final labeling result must satisfy all mutex constraints, which allows us to eliminate configurations that violate common sense physics laws like placing a floor above a night stand. Three classes of mutex constraints are proposed: global object co-occurrence constraint, relative height relationship constraint, and local support relationship constraint. We evaluate our approach on the NYU-Depth V2 dataset, which consists of 1449 cluttered indoor scenes, and also test generalization of our model trained on NYU-Depth V2 dataset directly on a recent SUN3D dataset without any new training. The experimental results show that we significantly outperform the state-of-the-art methods in scene labeling on both datasets.
Similar papers:
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • Semi-Supervised Normalized Cuts for Image Segmentation [pdf] - Selene E. Chew, Nathan D. Cahill
  • Constrained Convolutional Neural Networks for Weakly Supervised Segmentation [pdf] - Deepak Pathak, Philipp Krahenbuhl, Trevor Darrell
  • Semantic Component Analysis [pdf] - Calvin Murdock, Fernando De la Torre
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
Learning Large-Scale Automatic Image Colorization [pdf]
Aditya Deshpande, Jason Rock, David Forsyth

Abstract: We describe an automated method for image colorization that learns to colorize from examples. Our method exploits a LEARCH framework to train a quadratic objective function in the chromaticity maps, comparable to a Gaussian random field. The coefficients of the objective function are conditioned on image features, using a random forest. The objective function admits correlations on long spatial scales, and can control spatial error in the colorization of the image. Images are then colorized by minimizing this objective function. We demonstrate that our method strongly outperforms a natural baseline on large-scale experiments with images of real scenes using a demanding loss function. We demonstrate that learning a model that is conditioned on scene produces improved results. We show how to incorporate a desired color histogram into the objective function, and that doing so can lead to further improvements in results. data and prior knowledge to set up an optimization problem, which is solved to recover the desired representation. Rather than using domain knowledge to set up prior or likelihood terms, we train an optimization problem by requiring it to produce good colorizations of training data. Contributions: Our colorization method is learned from data, using a novel variant of LEARCH to balance pixelwise accuracy and spatial error. Comparable methods for training Gaussian Random Fields must impose positive definiteness constraints on the inverse covariance matrix, and enc
Similar papers:
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
  • Beyond White: Ground Truth Colors for Color Constancy Correction [pdf] - Dongliang Cheng, Brian Price, Scott Cohen, Michael S. Brown
  • Convolutional Color Constancy [pdf] - Jonathan T. Barron
Example-Based Modeling of Facial Texture From Deficient Data [pdf]
Arnaud Dessein, William A. P. Smith, Richard C. Wilson, Edwin R. Hancock

Abstract: We present an approach to modeling ear-to-ear, highquality texture from one or more partial views of a face with possibly poor resolution and noise. Our approach is example-based in that we reconstruct texture with patches from a database composed of previously seen faces. A 3D morphable model is used to establish shape correspondence between the observed data across views and training faces. The database is built on the mesh surface by segmenting it into uniform overlapping patches. Texture patches are selected by belief propagation so as to be consistent with neighbors and with observations in an appropriate image formation model. We also develop a variant that is insensitive to light and camera parameters, and incorporate soft symmetry constraints. We obtain textures of higher quality for degraded views as small as 10 pixels wide, than a standard model fitted to non-degraded data. We further show applications to super-resolution where we substantially improve quality compared to a state-of-the-art algorithm, and to texture completion where we fill in missing regions and remove facial clutter in a photorealistic manner.
Similar papers:
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Detailed Full-Body Reconstructions of Moving People From Monocular RGB-D Sequences [pdf] - Federica Bogo, Michael J. Black, Matthew Loper, Javier Romero
  • Robust Statistical Face Frontalization [pdf] - Christos Sagonas, Yannis Panagakis, Stefanos Zafeiriou, Maja Pantic
  • What Makes Tom Hanks Look Like Tom Hanks [pdf] - Supasorn Suwajanakorn, Steven M. Seitz, Ira Kemelmacher-Shlizerman
  • PatchMatch-Based Automatic Lattice Detection for Near-Regular Textures [pdf] - Siying Liu, Tian-Tsong Ng, Kalyan Sunkavalli, Minh N. Do, Eli Shechtman, Nathan Carr
Predicting Multiple Structured Visual Interpretations [pdf]
Debadeepta Dey, Varun Ramakrishna, Martial Hebert, J. Andrew Bagnell

Abstract: We present a simple approach for producing a small number of structured visual outputs which have high recall, for a variety of tasks including monocular pose estimation and semantic scene segmentation. Current state-of-the-art approaches learn a single model and modify inference procedures to produce a small number of diverse predictions. We take the alternate route of modifying the learning procedure to directly optimize for good, high recall sequences of structured-output predictors. Our approach introduces no new parameters, naturally learns diverse predictions and is not tied to any specific structured learning or inference procedure. We leverage recent advances in the contextual submodular maximization literature to learn a sequence of predictors and empirically demonstrate the simplicity and performance of our approach on multiple challenging vision tasks including achieving state-of-the-art results on multiple predictions for monocular pose-estimation and image foreground/background segmentation.
Similar papers:
  • Learning Ensembles of Potential Functions for Structured Prediction With Latent Variables [pdf] - Hossein Hajimirsadeghi, Greg Mori
  • Optimizing Expected Intersection-Over-Union With Candidate-Constrained CRFs [pdf] - Faruk Ahmed, Dany Tarlow, Dhruv Batra
  • Entropy-Based Latent Structured Output Prediction [pdf] - Diane Bouchacourt, Sebastian Nowozin, M. Pawan Kumar
  • Higher-Order Inference for Multi-Class Log-Supermodular Models [pdf] - Jian Zhang, Josip Djolonga, Andreas Krause
  • Inferring M-Best Diverse Labelings in a Single One [pdf] - Alexander Kirillov, Bogdan Savchynskyy, Dmitrij Schlesinger, Dmitry Vetrov, Carsten Rother
Contour Flow: Middle-Level Motion Estimation by Combining Motion Segmentation and Contour Alignment [pdf]
Huijun Di, Qingxuan Shi, Feng Lv, Ming Qin, Yao Lu

Abstract: Our goal is to estimate contour flow (the contour pairs with consistent point correspondence) from inconsistent contours extracted independently in two video frames. We formulate the contour flow estimation locally as a motion segmentation problem where motion patterns grouped from optical flow field are exploited for local correspondence measurement. To solve local ambiguities, contour flow estimation is further formulated globally as a contour alignment problem. We propose a novel two-staged strategy to obtain global consistent point correspondence under various contour transitions such as splitting, merging and branching. The goal of the first stage is to obtain possible accurate contour-to-contour alignments, and the second stage aims to make a consistent fusion of many partial alignments. Such a strategy can properly balance the accuracy and the consistency, which enables a middle-level motion representation to be constructed by just concatenating frame-by-frame contour flow estimation. Experiments prove the effectiveness of our method.
Similar papers:
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • Contour Box: Rejecting Object Proposals Without Explicit Closed Contours [pdf] - Cewu Lu, Shu Liu, Jiaya Jia, Chi-Keung Tang
  • Contour Guided Hierarchical Model for Shape Matching [pdf] - Yuanqi Su, Yuehu Liu, Bonan Cuan, Nanning Zheng
  • StereoSnakes: Contour Based Consistent Object Extraction For Stereo Images [pdf] - Ran Ju, Tongwei Ren, Gangshan Wu
  • Interactive Visual Hull Refinement for Specular and Transparent Object Surface Reconstruction [pdf] - Xinxin Zuo, Chao Du, Sen Wang, Jiangbin Zheng, Ruigang Yang
Convex Optimization With Abstract Linear Operators [pdf]
Steven Diamond, Stephen Boyd

Abstract: Linear Operators Steven Diamond and Stephen Boyd Dept. of Computer Science and Electrical Engineering, Stanford University {stevend2, boyd}@stanford.edu Abstract We introduce a convex optimization modeling framework that transforms a convex optimization problem expressed in a form natural and convenient for the user into an equivalent cone program in a way that preserves fast linear transforms in the original problem. By representing linear functions in the transformation process not as matrices, but as graphs that encode composition of abstract linear operators, we arrive at a matrix-free cone program, i.e., one whose data matrix is represented by an abstract linear operator and its adjoint. This cone program can then be solved by a matrix-free cone solver. By combining the matrix-free modeling framework and cone solver, we obtain a general method for efficiently solving convex optimization problems involving fast linear transforms.
Similar papers:
  • Multi-Image Matching via Fast Alternating Minimization [pdf] - Xiaowei Zhou, Menglong Zhu, Kostas Daniilidis
  • A Wavefront Marching Method for Solving the Eikonal Equation on Cartesian Grids [pdf] - Brais Cancela, Marcos Ortega, Manuel G. Penedo
  • Secrets of Matrix Factorization: Approximations, Numerics, Manifold Optimization and Random Restarts [pdf] - Je Hyeong Hong, Andrew Fitzgibbon
  • Structure From Motion Using Structure-Less Resection [pdf] - Enliang Zheng, Changchang Wu
  • Efficient Solution to the Epipolar Geometry for Radially Distorted Cameras [pdf] - Zuzana Kukelova, Jan Heller, Martin Bujnak, Andrew Fitzgibbon, Tomas Pajdla
kNN Hashing With Factorized Neighborhood Representation [pdf]
Kun Ding, Chunlei Huo, Bin Fan, Chunhong Pan

Abstract: Hashing is very effective for many tasks in reducing the processing time and in compressing massive databases. Although lots of approaches have been developed to learn data-dependent hash functions in recent years, how to learn hash functions to yield good performance with acceptable computational and memory cost is still a challenging problem. Based on the observation that retrieval precision is highly related to the kNN classification accuracy, this paper proposes a novel kNN-based supervised hashing method, which learns hash functions by directly maximizing the kNN accuracy of the Hamming-embedded training data. To make it scalable well to large problem, we propose a factorized neighborhood representation to parsimoniously model the neighborhood relationships inherent in training data. Considering that real-world data are often linearly inseparable, we further kernelize this basic model to improve its performance. As a result, the proposed method is able to learn accurate hashing functions with tolerable computation and storage cost. Experiments on four benchmarks demonstrate that our method outperforms the state-of-the-arts.
Similar papers:
  • Multi-View Complementary Hash Tables for Nearest Neighbor Search [pdf] - Xianglong Liu, Lei Huang, Cheng Deng, Jiwen Lu, Bo Lang
  • Learning Binary Codes for Maximum Inner Product Search [pdf] - Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen
  • An NMF Perspective on Binary Hashing [pdf] - Lopamudra Mukherjee, Sathya N. Ravi, Vamsi K. Ithapu, Tyler Holmes, Vikas Singh
  • Alternating Co-Quantization for Cross-Modal Hashing [pdf] - Go Irie, Hiroyuki Arai, Yukinobu Taniguchi
  • Adaptive Hashing for Fast Similarity Search [pdf] - Fatih Cakir, Stan Sclaroff
Probabilistic Label Relation Graphs With Ising Models [pdf]
Nan Ding, Jia Deng, Kevin P. Murphy, Hartmut Neven

Abstract: We consider classification problems in which the label space has structure. A common example is hierarchical label spaces, corresponding to the case where one label subsumes another (e.g., animal subsumes dog). But labels can also be mutually exclusive (e.g., dog vs cat) or unrelated (e.g., furry, carnivore). To jointly model hierarchy and exclusion relations, the notion of a HEX (hierarchy and exclusion) graph was introduced in [8]. This combined a conditional random field (CRF) with a deep neural network (DNN), resulting in state of the art results when applied to visual object classification problems where the training labels were drawn from different levels of the ImageNet hierarchy (e.g., an image might be labeled with the basic level category dog, rather than the more specific label husky). In this paper, we extend the HEX model to allow for soft or probabilistic relations between labels, which is useful when there is uncertainty about the relationship between two labels (e.g., an antelope is sort of furry, but not to the same degree as a grizzly bear). We call our new model pHEX, for probabilistic HEX. We show that the pHEX graph can be converted to an Ising model, which allows us to use existing off-the-shelf inference methods (in contrast to the HEX method, which needed specialized inference algorithms). Experimental results show significant improvements in a number of large-scale visual object classification tasks, outperforming the previous HEX model.
Similar papers:
  • Semi-Supervised Zero-Shot Classification With Label Representation Learning [pdf] - Xin Li, Yuhong Guo, Dale Schuurmans
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
  • Mining And-Or Graphs for Graph Matching and Object Discovery [pdf] - Quanshi Zhang, Ying Nian Wu, Song-Chun Zhu
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • ML-MG: Multi-Label Learning With Missing Labels Using a Mixed Graph [pdf] - Baoyuan Wu, Siwei Lyu, Bernard Ghanem
Unsupervised Visual Representation Learning by Context Prediction [pdf]
Carl Doersch, Abhinav Gupta, Alexei A. Efros

Abstract: This work explores the use of spatial context as a source of free and plentiful supervisory signal for training a rich visual representation. Given only a large, unlabeled image collection, we extract random pairs of patches from each image and train a convolutional neural net to predict the position of the second patch relative to the first. We argue that doing well on this task requires the model to learn to recognize objects and their parts. We demonstrate that the feature representation learned using this within-image context indeed captures visual similarity across images. For example, this representation allows us to perform unsupervised visual discovery of objects like cats, people, and even birds from the Pascal VOC 2011 detection dataset. Furthermore, we show that the learned ConvNet can be used in the RCNN framework [19] and provides a significant boost over a randomly-initialized ConvNet, resulting in state-of-theart performance among algorithms which use only Pascalprovided training set annotations.
Similar papers:
  • External Patch Prior Guided Internal Clustering for Image Denoising [pdf] - Fei Chen, Lei Zhang, Huimin Yu
  • BubbLeNet: Foveated Imaging for Visual Discovery [pdf] - Kevin Matzen, Noah Snavely
  • Local Convolutional Features With Unsupervised Training for Image Retrieval [pdf] - Mattis Paulin, Matthijs Douze, Zaid Harchaoui, Julien Mairal, Florent Perronin, Cordelia Schmid
  • Harvesting Discriminative Meta Objects With Deep CNN Features for Scene Classification [pdf] - Ruobing Wu, Baoyuan Wang, Wenping Wang, Yizhou Yu
  • Unsupervised Learning of Visual Representations Using Videos [pdf] - Xiaolong Wang, Abhinav Gupta
Parsimonious Labeling [pdf]
Puneet K. Dokania, M. Pawan Kumar

Abstract: We propose a new family of discrete energy minimization problems, which we call parsimonious labeling. Our energy function consists of unary potentials and high-order clique potentials. While the unary potentials are arbitrary, the clique potentials are proportional to the diversity of the set of unique labels assigned to the clique. Intuitively, our energy function encourages the labeling to be parsimonious, that is, use as few labels as possible. This in turn allows us to capture useful cues for important computer vision applications such as stereo correspondence and image denoising. Furthermore, we propose an efficient graph-cuts based algorithm for the parsimonious labeling problem that provides strong theoretical guarantees on the quality of the solution. Our algorithm consists of three steps. First, we approximate a given diversity using a mixture of a novel hierarchical P n Potts model. Second, we use a divide-andconquer approach for each mixture component, where each subproblem is solved using an efficient -expansion algorithm. This provides us with a small number of putative labelings, one for each mixture component. Third, we choose the best putative labeling in terms of the energy value. Using both synthetic and standard real datasets, we show that our algorithm significantly outperforms other graph-cuts based approaches.
Similar papers:
  • Higher-Order CRF Structural Segmentation of 3D Reconstructed Surfaces [pdf] - Jingbo Liu, Jinglu Wang, Tian Fang, Chiew-Lan Tai, Long Quan
  • A Multiscale Variable-Grouping Framework for MRF Energy Minimization [pdf] - Omer Meir, Meirav Galun, Stav Yagev, Ronen Basri, Irad Yavneh
  • Semantic Segmentation With Object Clique Potential [pdf] - Xiaojuan Qi, Jianping Shi, Shu Liu, Renjie Liao, Jiaya Jia
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
  • Inferring M-Best Diverse Labelings in a Single One [pdf] - Alexander Kirillov, Bogdan Savchynskyy, Dmitrij Schlesinger, Dmitry Vetrov, Carsten Rother
Compression Artifacts Reduction by a Deep Convolutional Network [pdf]
Chao Dong, Yubin Deng, Chen Change Loy, Xiaoou Tang

Abstract: Lossy compression introduces complex compression artifacts, particularly the blocking artifacts, ringing effects and blurring. Existing algorithms either focus on removing blocking artifacts and produce blurred output, or restores sharpened images that are accompanied with ringing effects. Inspired by the deep convolutional networks (DCN) on super-resolution [5], we formulate a compact and efcient network for seamless attenuation of different compression artifacts. We also demonstrate that a deeper model can be effectively trained with the features learned in a shallow network. Following a similar easy to hard idea, we systematically investigate several practical transfer settings and show the effectiveness of transfer learning in low-level vision problems. Our method shows superior performance than the state-of-the-arts both on the benchmark datasets and the real-world use case (i.e. Twitter).
Similar papers:
  • Improving Image Restoration With Soft-Rounding [pdf] - Xing Mei, Honggang Qi, Bao-Gang Hu, Siwei Lyu
  • Structured Feature Selection [pdf] - Tian Gao, Ziheng Wang, Qiang Ji
  • Deep Networks for Image Super-Resolution With Sparse Prior [pdf] - Zhaowen Wang, Ding Liu, Jianchao Yang, Wei Han, Thomas Huang
  • An Exploration of Parameter Redundancy in Deep Networks With Circulant Projections [pdf] - Yu Cheng, Felix X. Yu, Rogerio S. Feris, Sanjiv Kumar, Alok Choudhary, Shi-Fu Chang
  • Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification [pdf] - Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun
Low-Rank Tensor Approximation With Laplacian Scale Mixture Modeling for Multiframe Image Denoising [pdf]
Weisheng Dong, Guangyu Li, Guangming Shi, Xin Li, Yi Ma

Abstract: Patch-based low-rank models have shown effective in exploiting spatial redundancy of natural images especially for the application of image denoising. However, twodimensional low-rank model can not fully exploit the spatiotemporal correlation in larger data sets such as multispectral images and 3D MRIs. In this work, we propose a novel low-rank tensor approximation framework with Laplacian Scale Mixture (LSM) modeling for multi-frame image denoising. First, similar 3D patches are grouped to form a tensor of d-order and high-order Singular Value Decomposition (HOSVD) is applied to the grouped tensor. Then the task of multiframe image denoising is formulated as a Maximum A Posterior (MAP) estimation problem with the LSM prior for tensor coefficients. Both unknown sparse coefficients and hidden LSM parameters can be efficiently estimated by the method of alternating optimization. Specifically, we have derived closed-form solutions for both subproblems. Experimental results on spectral and dynamic MRI images show that the proposed algorithm can better preserve the sharpness of important image structures and outperform several existing state-of-the-art multiframe denoising methods (e.g., BM4D and tensor dictionary learning).
Similar papers:
  • Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf] - Yuhui Quan, Yan Huang, Hui Ji
  • External Patch Prior Guided Internal Clustering for Image Denoising [pdf] - Fei Chen, Lei Zhang, Huimin Yu
  • Patch Group Based Nonlocal Self-Similarity Prior Learning for Image Denoising [pdf] - Jun Xu, Lei Zhang, Wangmeng Zuo, David Zhang, Xiangchu Feng
  • Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf] - Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • A Novel Sparsity Measure for Tensor Recovery [pdf] - Qian Zhao, Deyu Meng, Xu Kong, Qi Xie, Wenfei Cao, Yao Wang, Zongben Xu
3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf]
Matea Donlic, Tomislav Petkovic, Tomislav Pribanic

Abstract: A novel structured light method for color 3D surface profilometry is proposed. The proposed method does not require color calibration of a camera-projector pair and may be used for reconstruction of both dynamic and static scenes. The method uses a structured light pattern that is a combination of a De Bruijn color sequence and of a sinusoidal fringe. For dynamic scenes a Hessian ridge detector and a Gaussian mixture model are combined to extract stripe centers and to identify color. Stripes are then uniquely identified using dynamic programming based on the Smith-Waterman algorithm and a De Bruijn window property. For static scenes phase-shifting and De Bruijn window property are combined to obtain a high accuracy reconstruction. We have tested the proposed method on multiple objects with challenging surfaces and different albedos that demonstrate usability and robustness of the method.
Similar papers:
  • General Dynamic Scene Reconstruction From Multiple View Video [pdf] - Armin Mustafa, Hansung Kim, Jean-Yves Guillemaut, Adrian Hilton
  • Nighttime Haze Removal With Glow and Multiple Light Colors [pdf] - Yu Li, Robby T. Tan, Michael S. Brown
  • Convolutional Color Constancy [pdf] - Jonathan T. Barron
  • Beyond White: Ground Truth Colors for Color Constancy Correction [pdf] - Dongliang Cheng, Brian Price, Scott Cohen, Michael S. Brown
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
Point Triangulation Through Polyhedron Collapse Using the l[?] Norm [pdf]
Simon Donne, Bart Goossens, Wilfried Philips

Abstract: Multi-camera triangulation of feature points based on a minimisation of the overall 2 reprojection error can get stuck in suboptimal local minima or require slow global optimisation. For this reason, researchers have proposed optimising the norm of the 2 single view reprojection errors, which avoids the problem of local minima entirely. In this paper we present a novel method for triangulation that minimizes the norm of the reprojection errors: this apparently small difference leads to a much faster but equally accurate solution which is related to the MLE under the assumption of uniform noise. The proposed method adopts a new optimisation strategy based on solving simple quadratic equations. This stands in contrast with the fastest existing methods, which solve a sequence of more complex auxiliary Linear Programming or Second Order Cone Problems. The proposed algorithm performs well: for triangulation, it achieves the same accuracy as existing techniques while executing faster and being straightforward to implement.
Similar papers:
  • The Joint Image Handbook [pdf] - Matthew Trager, Martial Hebert, Jean Ponce
  • Low-Rank Matrix Factorization Under General Mixture Noise Distributions [pdf] - Xiangyong Cao, Yang Chen, Qian Zhao, Deyu Meng, Yao Wang, Dong Wang, Zongben Xu
  • Towards Pointless Structure From Motion: 3D Reconstruction and Camera Parameters From General 3D Curves [pdf] - Irina Nurutdinova, Andrew Fitzgibbon
  • Active Transfer Learning With Zero-Shot Priors: Reusing Past Datasets for Future Tasks [pdf] - Efstratios Gavves, Thomas Mensink, Tatiana Tommasi, Cees G. M. Snoek, Tinne Tuytelaars
  • Multi-Class Multi-Annotator Active Learning With Robust Gaussian Process for Visual Recognition [pdf] - Chengjiang Long, Gang Hua
FlowNet: Learning Optical Flow With Convolutional Networks [pdf]
Alexey Dosovitskiy, Philipp Fischer, Eddy Ilg, Philip Hausser, Caner Hazirbas, Vladimir Golkov, Patrick van der Smagt, Daniel Cremers, Thomas Brox

Abstract: Convolutional neural networks (CNNs) have recently been very successful in a variety of computer vision tasks, especially on those linked to recognition. Optical flow estimation has not been among the tasks CNNs succeeded at. In this paper we construct CNNs which are capable of solving the optical flow estimation problem as a supervised learning task. We propose and compare two architectures: a generic architecture and another one including a layer that correlates feature vectors at different image locations. Since existing ground truth data sets are not sufficiently large to train a CNN, we generate a large synthetic Flying Chairs dataset. We show that networks trained on this unrealistic data still generalize very well to existing datasets such as Sintel and KITTI, achieving competitive accuracy at frame rates of 5 to 10 fps.
Similar papers:
  • Face Flow [pdf] - Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou
  • Dense Optical Flow Prediction From a Static Image [pdf] - Jacob Walker, Abhinav Gupta, Martial Hebert
  • Oriented Light-Field Windows for Scene Flow [pdf] - Pratul P. Srinivasan, Michael W. Tao, Ren Ng, Ravi Ramamoorthi
  • Large Displacement 3D Scene Flow With Occlusion Reasoning [pdf] - Andrei Zanfir, Cristian Sminchisescu
  • Flow Fields: Dense Correspondence Fields for Highly Accurate Large Displacement Optical Flow Estimation [pdf] - Christian Bailer, Bertram Taetz, Didier Stricker
What Makes an Object Memorable? [pdf]
Rachit Dubey, Joshua Peterson, Aditya Khosla, Ming-Hsuan Yang, Bernard Ghanem

Abstract: Recent studies on image memorability have shed light on what distinguishes the memorability of different images and the intrinsic and extrinsic properties that make those images memorable. However, a clear understanding of the memorability of specific objects inside an image remains elusive. In this paper, we provide the first attempt to answer the question: what exactly is remembered about an image? We augment both the images and object segmentations from the PASCAL-S dataset with ground truth memorability scores and shed light on the various factors and properties that make an object memorable (or forgettable) to humans. We analyze various visual factors that may influence object memorability (e.g. color, visual saliency, and object categories). We also study the correlation between object and image memorability and find that image memorability is greatly affected by the memorability of its most memorable object. Lastly, we explore the effectiveness of deep learning and other computational approaches in predicting object memorability in images. Our efforts offer a deeper understanding of memorability in general thereby opening up avenues for a wide variety of applications.
Similar papers:
  • HARF: Hierarchy-Associated Rich Features for Salient Object Detection [pdf] - Wenbin Zou, Nikos Komodakis
  • Understanding and Predicting Image Memorability at a Large Scale [pdf] - Aditya Khosla, Akhil S. Raju, Antonio Torralba, Aude Oliva
  • Cluster-Based Point Set Saliency [pdf] - Flora Ponjou Tasse, Jiri Kosinka, Neil Dodgson
  • A Data-Driven Metric for Comprehensive Evaluation of Saliency Models [pdf] - Jia Li, Changqun Xia, Yafei Song, Shu Fang, Xiaowu Chen
  • SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf] - Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao
MANTRA: Minimum Maximum Latent Structural SVM for Image Classification and Ranking [pdf]
Thibaut Durand, Nicolas Thome, Matthieu Cord

Abstract: In this work, we propose a novel Weakly Supervised Learning (WSL) framework dedicated to learn discriminative part detectors from images annotated with a global label. Our WSL method encompasses three main contributions. Firstly, we introduce a new structured output latent variable model, Minimum mAximum lateNt sTRucturAl SVM (MANTRA), which prediction relies on a pair of latent variables: h+ (resp. h ) provides positive (resp. negative) evidence for a given output y. Secondly, we instantiate MANTRA for two different visual recognition tasks: multiclass classification and ranking. For ranking, we propose efficient solutions to exactly solve the inference and the lossaugmented problems. Finally, extensive experiments highlight the relevance of the proposed method: MANTRA outperforms state-of-the art results on five different datasets.
Similar papers:
  • A Nonparametric Bayesian Approach Toward Stacked Convolutional Independent Component Analysis [pdf] - Sotirios P. Chatzis, Dimitrios Kosmopoulos
  • Cutting Edge: Soft Correspondences in Multimodal Scene Parsing [pdf] - Sarah Taghavi Namin, Mohammad Najafi, Mathieu Salzmann, Lars Petersson
  • Similarity Gaussian Process Latent Variable Model for Multi-Modal Data Analysis [pdf] - Guoli Song, Shuhui Wang, Qingming Huang, Qi Tian
  • Learning to Rank Based on Subsequences [pdf] - Basura Fernando, Efstratios Gavves, Damien Muselet, Tinne Tuytelaars
  • Entropy-Based Latent Structured Output Prediction [pdf] - Diane Bouchacourt, Sebastian Nowozin, M. Pawan Kumar
Predicting Depth, Surface Normals and Semantic Labels With a Common Multi-Scale Convolutional Architecture [pdf]
David Eigen, Rob Fergus

Abstract: In this paper we address three different computer vision tasks using a single multiscale convolutional network architecture: depth prediction, surface normal estimation, and semantic labeling. The network that we develop is able to adapt naturally to each task using only small modifications, regressing from the input image to the output map directly. Our method progressively refines predictions using a sequence of scales, and captures many image details without any superpixels or low-level segmentation. We achieve state-of-the-art performance on benchmarks for all three tasks.
Similar papers:
  • Learning Deconvolution Network for Semantic Segmentation [pdf] - Hyeonwoo Noh, Seunghoon Hong, Bohyung Han
  • Joint Object and Part Segmentation Using Deep Learned Potentials [pdf] - Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille
  • Monocular Object Instance Segmentation and Depth Ordering With CNNs [pdf] - Ziyu Zhang, Alexander G. Schwing, Sanja Fidler, Raquel Urtasun
  • MMSS: Multi-Modal Sharable and Specific Feature Learning for RGB-D Object Recognition [pdf] - Anran Wang, Jianfei Cai, Jiwen Lu, Tat-Jen Cham
  • Polarized 3D: High-Quality Depth Sensing With Polarization Cues [pdf] - Achuta Kadambi, Vage Taamazyan, Boxin Shi, Ramesh Raskar
Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf]
Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic

Abstract: We propose a novel multi-conditional latent variable model for simultaneous facial feature fusion and detection of facial action units. In our approach we exploit the structure-discovery capabilities of generative models such as Gaussian processes, and the discriminative power of classifiers such as logistic function. This leads to superior performance compared to existing classifiers for the target task that exploit either the discriminative or generative property, but not both. The model learning is performed via an efficient, newly proposed Bayesian learning strategy based on Monte Carlo sampling. Consequently, the learned model is robust to data overfitting, regardless of the number of both input features and jointly estimated facial action units. Extensive qualitative and quantitative experimental evaluations are performed on three publicly available datasets (CK+, Shoulder-pain and DISFA). We show that the proposed model outperforms the state-of-theart methods for the target task on (i) feature fusion, and (ii) multiple facial action unit detection.
Similar papers:
  • Joint Fine-Tuning in Deep Neural Networks for Facial Expression Recognition [pdf] - Heechul Jung, Sihaeng Lee, Junho Yim, Sunjeong Park, Junmo Kim
  • A Nonparametric Bayesian Approach Toward Stacked Convolutional Independent Component Analysis [pdf] - Sotirios P. Chatzis, Dimitrios Kosmopoulos
  • From Emotions to Action Units With Hidden and Semi-Hidden-Task Learning [pdf] - Adria Ruiz, Joost Van de Weijer, Xavier Binefa
  • Learning to Transfer: Transferring Latent Task Structures and Its Application to Person-Specific Facial Action Unit Detection [pdf] - Timur Almaev, Brais Martinez, Michel Valstar
  • Similarity Gaussian Process Latent Variable Model for Multi-Modal Data Analysis [pdf] - Guoli Song, Shuhui Wang, Qingming Huang, Qi Tian
Fine-Grained Change Detection of Misaligned Scenes With Varied Illuminations [pdf]
Wei Feng, Fei-Peng Tian, Qian Zhang, Nan Zhang, Liang Wan, Jizhou Sun

Abstract: 3 1 Detecting fine-grained subtle changes among a scene is critically important in practice. Previous change detection methods, focusing on detecting large-scale significant changes, cannot do this well. This paper proposes a feasible end-to-end approach to this challenging problem. We start from active camera relocation that quickly relocates camera to nearly the same pose and position of the last time observation. To guarantee detection sensitivity and accuracy of minute changes, in an observation, we capture a group of images under multiple illuminations, which need only to be roughly aligned to the last time lighting conditions. Given two times observations, we formulate fine-grained change detection as a joint optimization problem of three related factors, i.e., normal-aware lighting difference, camera geometry correction flow, and real scene change mask. We solve the three factors in a coarse-to-fine manner and achieve reliable change decision by rank minimization. We build three real-world datasets to benchmark fine-grained change detection of misaligned scenes under varied multiple lighting conditions. Extensive experiments show the superior performance of our approach over state-of-the-art change detection methods and its ability to distinguish real scene changes from false ones caused by lighting variations.
Similar papers:
  • Robust RGB-D Odometry Using Point and Line Features [pdf] - Yan Lu, Dezhen Song
  • Robust Heart Rate Measurement From Video Using Select Random Patches [pdf] - Antony Lam, Yoshinori Kuno
  • A Comprehensive Multi-Illuminant Dataset for Benchmarking of the Intrinsic Image Algorithms [pdf] - Shida Beigpour, Andreas Kolb, Sven Kunz
  • Beyond White: Ground Truth Colors for Color Constancy Correction [pdf] - Dongliang Cheng, Brian Price, Scott Cohen, Michael S. Brown
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
Learning The Structure of Deep Convolutional Networks [pdf]
Jiashi Feng, Trevor Darrell

Abstract: Training data In this work, we develop a novel method for automatically learning aspects of the structure of a deep model in order to improve its performance, especially when labeled training data are scarce. We propose a new convolutional neural network model with the Indian Buffet Process (IBP) prior, termed ibpCNN. The ibpCNN automatically adapts its structure to provided training data, achieves an optimal balance among model complexity, data delity and training loss, and thus offers better generalization performance. The proposed ibpCNN captures complex data distribution in an unsupervised generative way. Therefore, ibpCNN can exploit unlabeled data which can be collected at low cost to learn its structure. After determining the structure, ibpCNN further learns its parameters according to specied tasks, in an end-to-end fashion, and produces discriminative yet compact representations. We evaluate the performance of ibpCNN, on fully- and semi-supervised image classication tasks; ibpCNN surpasses standard CNN models on benchmark datasets with much smaller size and higher efciency.
Similar papers:
  • Multi-Scale Recognition With DAG-CNNs [pdf] - Songfan Yang, Deva Ramanan
  • Fast R-CNN [pdf] - Ross Girshick
  • Deep Fried Convnets [pdf] - Zichao Yang, Marcin Moczulski, Misha Denil, Nando de Freitas, Alex Smola, Le Song, Ziyu Wang
  • Matrix Backpropagation for Deep Networks With Structured Layers [pdf] - Catalin Ionescu, Orestis Vantzos, Cristian Sminchisescu
  • An Exploration of Parameter Redundancy in Deep Networks With Circulant Projections [pdf] - Yu Cheng, Felix X. Yu, Rogerio S. Feris, Sanjiv Kumar, Alok Choudhary, Shi-Fu Chang
Continuous Pose Estimation With a Spatial Ensemble of Fisher Regressors [pdf]
Michele Fenzi, Laura Leal-Taixe, Jorn Ostermann, Tinne Tuytelaars

Abstract: In this paper, we treat the problem of continuous pose estimation for object categories as a regression problem on the basis of only 2D training information. While regression is a natural framework for continuous problems, regression methods so far achieved inferior results with respect to 3D-based and 2D-based classification-and-refinement approaches. This may be attributed to their weakness to high intra-class variability as well as to noisy matching procedures and lack of geometrical constraints. We propose to apply regression to Fisher-encoded vectors computed from large cells by learning an array of Fisher regressors. Fisher encoding makes our algorithm flexible to variations in class appearance, while the array structure permits to indirectly introduce spatial context information in the approach. We formulate our problem as a MAP inference problem, where the likelihood function is composed of a generative term based on the prediction error generated by the ensemble of Fisher regressors as well as a discriminative term based on SVM classifiers. We test our algorithm on three publicly available datasets that envisage several difficulties, such as high intra-class variability, truncations, occlusions, and motion blur, obtaining state-of-the-art results.
Similar papers:
  • Render for CNN: Viewpoint Estimation in Images Using CNNs Trained With Rendered 3D Model Views [pdf] - Hao Su, Charles R. Qi, Yangyan Li, Leonidas J. Guibas
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • Unsupervised Generation of a Viewpoint Annotated Car Dataset From Videos [pdf] - Nima Sedaghat, Thomas Brox
  • Pose Induction for Novel Object Categories [pdf] - Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • Real-Time Pose Estimation Piggybacked on Object Detection [pdf] - Roman Juranek, Adam Herout, Marketa Dubska, Pavel Zemcik
Learning to Rank Based on Subsequences [pdf]
Basura Fernando, Efstratios Gavves, Damien Muselet, Tinne Tuytelaars

Abstract: We present a supervised learning to rank algorithm that effectively orders images by exploiting the structure in image sequences. Most often in the supervised learning to rank literature, ranking is approached either by analyzing pairs of images or by optimizing a list-wise surrogate loss function on full sequences. In this work we propose MidRank, which learns from moderately sized sub-sequences instead. These sub-sequences contain useful structural ranking information that leads to better learnability during training and better generalization during testing. By exploiting sub-sequences, the proposed MidRank improves ranking accuracy considerably on an extensive array of image ranking applications and datasets.
Similar papers:
  • Just Noticeable Differences in Visual Attributes [pdf] - Aron Yu, Kristen Grauman
  • Infinite Feature Selection [pdf] - Giorgio Roffo, Simone Melzi, Marco Cristani
  • Person Re-Identification Ranking Optimisation by Discriminant Context Information Analysis [pdf] - Jorge Garcia, Niki Martinel, Christian Micheloni, Alfredo Gardel
  • Attribute-Graph: A Graph Based Approach to Image Ranking [pdf] - Nikita Prabhu, R. Venkatesh Babu
  • MANTRA: Minimum Maximum Latent Structural SVM for Image Classification and Ranking [pdf] - Thibaut Durand, Nicolas Thome, Matthieu Cord
Variational Depth Superresolution Using Example-Based Edge Representations [pdf]
David Ferstl, Matthias Ruther, Horst Bischof

Abstract: In this paper we propose a novel method for depth image superresolution which combines recent advances in example based upsampling with variational superresolution based on a known blur kernel. Most traditional depth superresolution approaches try to use additional high resolution intensity images as guidance for superresolution. In our method we learn a dictionary of edge priors from an external database of high and low resolution examples. In a novel variational sparse coding approach this dictionary is used to infer strong edge priors. Additionally to the traditional sparse coding constraints the difference in the overlap of neighboring edge patches is minimized in our optimization. These edge priors are used in a novel variational superresolution as anisotropic guidance of a higher order regularization. Both the sparse coding and the variational superresolution of the depth are solved based on the primaldual formulation. In an exhaustive numerical and visual evaluation we show that our method clearly outperforms existing approaches on multiple real and synthetic datasets. depth images from nearby locations [23]. However, in previous works it has been shown that it is possible to superresolve intensity images either by interpolation and deblurring with a known Point Spread Function (PSF) [26] or by previously learned relationships between low and high resolution image examples, that are stored in dictionaries [30]. On the one side, superresolution (SR) approaches based on
Similar papers:
  • Learning Parametric Distributions for Image Super-Resolution: Where Patch Matching Meets Sparse Coding [pdf] - Yongbo Li, Weisheng Dong, Guangming Shi, Xuemei Xie
  • Deep Networks for Image Super-Resolution With Sparse Prior [pdf] - Zhaowen Wang, Ding Liu, Jianchao Yang, Wei Han, Thomas Huang
  • Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf] - Yuhui Quan, Yan Huang, Hui Ji
  • Convolutional Sparse Coding for Image Super-Resolution [pdf] - Shuhang Gu, Wangmeng Zuo, Qi Xie, Deyu Meng, Xiangchu Feng, Lei Zhang
  • Conformal and Low-Rank Sparse Representation for Image Restoration [pdf] - Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng
POP Image Fusion - Derivative Domain Image Fusion Without Reintegration [pdf]
Graham D. Finlayson, Alex E. Hayes

Abstract: There are many applications where multiple images are fused to form a single summary greyscale or colour output, including computational photography (e.g. RGB-NIR), diffusion tensor imaging (medical), and remote sensing. Often, and intuitively, image fusion is carried out in the derivative domain. Here, a new composite fused derivative is found that best accounts for the detail across all images and then the resulting gradient field is reintegrated. However, the reintegration step generally hallucinates new detail (not appearing in any of the input image bands) including halo and bending artifacts. In this paper we avoid these hallucinated details by avoiding the reintegration step. Our work builds directly on the work of Socolinsky and Wolff who derive their equivalent gradient field from the per-pixel Di Zenzo structure tensor which is defined as the inner product of the image Jacobian. We show that the x- and y- derivatives of the projection of the original image onto the Principal characteristic vector of the Outer Product (POP) of the Jacobian generates the same equivalent gradient field. In so doing, we have derived a fused image that has the derivative structure we seek. Of course, this projection will be meaningful only where the Jacobian has non-zero derivatives, so we diffuse the projection directions using a bilateral filter before we calculate the fused image. The resulting POP fused image has maximal fused detail but avoids hallucinated artifacts. Experiments dem
Similar papers:
  • Query Adaptive Similarity Measure for RGB-D Object Recognition [pdf] - Yanhua Cheng, Rui Cai, Chi Zhang, Zhiwei Li, Xin Zhao, Kaiqi Huang, Yong Rui
  • Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf] - Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic
  • Fast and Effective L0 Gradient Minimization by Region Fusion [pdf] - Rang M. H. Nguyen, Michael S. Brown
  • Multiple Feature Fusion via Weighted Entropy for Visual Tracking [pdf] - Lin Ma, Jiwen Lu, Jianjiang Feng, Jie Zhou
  • Pan-Sharpening With a Hyper-Laplacian Penalty [pdf] - Yiyong Jiang, Xinghao Ding, Delu Zeng, Yue Huang, John Paisley
Single Image 3D Without a Single 3D Image [pdf]
David F. Fouhey, Wajahat Hussain, Abhinav Gupta, Martial Hebert

Abstract: Image Do we really need 3D labels in order to learn how to predict 3D? In this paper, we show that one can learn a mapping from appearance to 3D properties without ever seeing a single explicit 3D label. Rather than use explicit supervision, we use the regularity of indoor scenes to learn the mapping in a completely unsupervised manner. We demonstrate this on both a standard 3D scene understanding dataset as well as Internet images for which 3D is unavailable, precluding supervised learning. Despite never seeing a 3D label, our method produces competitive results. 3D Structure Style
Similar papers:
  • Harvesting Discriminative Meta Objects With Deep CNN Features for Scene Classification [pdf] - Ruobing Wu, Baoyuan Wang, Wenping Wang, Yizhou Yu
  • Deep Multi-Patch Aggregation Network for Image Style, Aesthetics, and Quality Estimation [pdf] - Xin Lu, Zhe Lin, Xiaohui Shen, Radomir Mech, James Z. Wang
  • Understanding Deep Features With Computer-Generated Imagery [pdf] - Mathieu Aubry, Bryan C. Russell
  • BubbLeNet: Foveated Imaging for Visual Discovery [pdf] - Kevin Matzen, Noah Snavely
  • Learning Visual Clothing Style With Heterogeneous Dyadic Co-Occurrences [pdf] - Andreas Veit, Balazs Kovacs, Sean Bell, Julian McAuley, Kavita Bala, Serge Belongie
Recurrent Network Models for Human Dynamics [pdf]
Katerina Fragkiadaki, Sergey Levine, Panna Felsen, Jitendra Malik

Abstract: We propose the Encoder-Recurrent-Decoder (ERD) model for recognition and prediction of human body pose in videos and motion capture. The ERD model is a recurrent neural network that incorporates nonlinear encoder and decoder networks before and after recurrent layers. We test instantiations of ERD architectures in the tasks of motion capture (mocap) generation, body pose labeling and body pose forecasting in videos. Our model handles mocap training data across multiple subjects and activity domains, and synthesizes novel motions while avoiding drifting for long periods of time. For human pose labeling, ERD outperforms a per frame body part detector by resolving left-right body part confusions. For video pose forecasting, ERD predicts body joint displacements across a temporal horizon of 400ms and outperforms a first order motion model based on optical flow. ERDs extend previous Long Short Term Memory (LSTM) models in the literature to jointly learn representations and their dynamics. Our experiments show such representation learning is crucial for both labeling and prediction in space-time. We find this is a distinguishing feature between the spatio-temporal visual domain in comparison to 1D text, speech or handwriting, where straightforward hard coded representations have shown excellent results when directly combined with recurrent units [31] .
Similar papers:
  • Flowing ConvNets for Human Pose Estimation in Videos [pdf] - Tomas Pfister, James Charles, Andrew Zisserman
  • Dense Optical Flow Prediction From a Static Image [pdf] - Jacob Walker, Abhinav Gupta, Martial Hebert
  • Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf] - Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang
  • Uncovering Interactions and Interactors: Joint Estimation of Head, Body Orientation and F-Formations From Surveillance Videos [pdf] - Elisa Ricci, Jagannadan Varadarajan, Ramanathan Subramanian, Samuel Rota Bulo, Narendra Ahuja, Oswald Lanz
  • Human Pose Estimation in Videos [pdf] - Dong Zhang, Mubarak Shah
Highly-Expressive Spaces of Well-Behaved Transformations: Keeping It Simple [pdf]
Oren Freifeld, Soren Hauberg, Kayhan Batmanghelich, John W. Fisher III

Abstract: v (x) n n We propose novel finite-dimensional spaces of R R transformations, n {1, 2, 3}, derived from (continuouslydefined) parametric stationary velocity fields. Particularly, we obtain these transformations, which are diffeomorphisms, by fast and highly-accurate integration of continuous piecewise-affine velocity fields; we also provide an exact solution for n = 1. The simple-yet-highly-expressive proposed representation handles optional constraints (e.g., volume preservation) easily and supports convenient modeling choices and rapid likelihood evaluations (facilitating tractable inference over latent transformations). Its applications include, but are not limited to: unconstrained optimization over monotonic functions; modeling cumulative distribution functions or histograms; time warping; image registration; landmark-based warping; real-time diffeomorphic image editing. Our code is available at https://github.com/freifeld/cpabDiffeo
Similar papers:
  • Convex Optimization With Abstract Linear Operators [pdf] - Steven Diamond, Stephen Boyd
  • Shell PCA: Statistical Shape Modelling in Shell Space [pdf] - Chao Zhang, Behrend Heeren, Martin Rumpf, William A. P. Smith
  • Realtime Edge-Based Visual Odometry for a Monocular Camera [pdf] - Juan Jose Tarrio, Sol Pedre
  • Compositional Hierarchical Representation of Shape Manifolds for Classification of Non-Manifold Shapes [pdf] - Mete Ozay, Umit Rusen Aktas, Jeremy L. Wyatt, Ales Leonardis
  • Recursive Frechet Mean Computation on the Grassmannian and its Applications to Computer Vision [pdf] - Rudrasis Chakraborty, Baba C. Vemuri
Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf]
Ying Fu, Antony Lam, Imari Sato, Yoichi Sato

Abstract: Hyperspectral imaging is beneficial in a diverse range of applications from diagnostic medicine, to agriculture, to surveillance to name a few. However, hyperspectral images often times suffer from degradation due to the limited light, which introduces noise into the imaging process. In this paper, we propose an effective model for hyperspectral image (HSI) denoising that considers underlying characteristics of HSIs: sparsity across the spatial-spectral domain, high correlation across spectra, and non-local self-similarity over space. We first exploit high correlation across spectra and non-local self-similarity over space in the noisy HSI to learn an adaptive spatial-spectral dictionary. Then, we employ the local and non-local sparsity of the HSI under the learned spatial-spectral dictionary to design an HSI denoising model, which can be effectively solved by an iterative numerical algorithm with parameters that are adaptively adjusted for different clusters and different noise levels. Experimental results on HSI denoising show that the proposed method can provide substantial improvements over the current state-of-the-art HSI denoising methods in terms of both objective metric and subjective visual quality.
Similar papers:
  • Conformal and Low-Rank Sparse Representation for Image Restoration [pdf] - Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng
  • RGB-Guided Hyperspectral Image Upsampling [pdf] - Hyeokhyen Kwon, Yu-Wing Tai
  • Patch Group Based Nonlocal Self-Similarity Prior Learning for Image Denoising [pdf] - Jun Xu, Lei Zhang, Wangmeng Zuo, David Zhang, Xiangchu Feng
  • Hyperspectral Super-Resolution by Coupled Spectral Unmixing [pdf] - Charis Lanaras, Emmanuel Baltsavias, Konrad Schindler
  • Low-Rank Tensor Approximation With Laplacian Scale Mixture Modeling for Multiframe Image Denoising [pdf] - Weisheng Dong, Guangyu Li, Guangming Shi, Xin Li, Yi Ma
Beyond Tree Structure Models: A New Occlusion Aware Graphical Model for Human Pose Estimation [pdf]
Lianrui Fu, Junge Zhang, Kaiqi Huang

Abstract: Occlusion is a main challenge for human pose estimation, which is largely ignored in popular tree structure models. The tree structure model is simple and convenient for exact inference, but short in modeling the occlusion coherence especially in the case of self-occlusion. We propose an occlusion aware graphical model which is able to model both self-occlusion and occlusion by the other objects simultaneously. The proposed model structure can encodes the interactions between human body parts and objects, and hence enables it to learn occlusion coherence from data discriminatively. We evaluate our model on several public benchmarks for human pose estimation including challenging subsets featuring significant occlusion. The experimental results show that our method obtains comparable accuracy with the state-of-the-arts, and is robust to occlusion for 2D human pose estimation.
Similar papers:
  • Robust Facial Landmark Detection Under Significant Head Poses and Occlusion [pdf] - Yue Wu, Qiang Ji
  • Deep Learning Strong Parts for Pedestrian Detection [pdf] - Yonglong Tian, Ping Luo, Xiaogang Wang, Xiaoou Tang
  • Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf] - Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang
  • Uncovering Interactions and Interactors: Joint Estimation of Head, Body Orientation and F-Formations From Surveillance Videos [pdf] - Elisa Ricci, Jagannadan Varadarajan, Ramanathan Subramanian, Samuel Rota Bulo, Narendra Ahuja, Oswald Lanz
  • Human Pose Estimation in Videos [pdf] - Dong Zhang, Mubarak Shah
Relaxing From Vocabulary: Robust Weakly-Supervised Deep Learning for Vocabulary-Free Image Tagging [pdf]
Jianlong Fu, Yue Wu, Tao Mei, Jinqiao Wang, Hanqing Lu, Yong Rui

Abstract: Input: Web Images X Input: Label Y bird
Similar papers:
  • Look and Think Twice: Capturing Top-Down Visual Attention With Feedback Convolutional Neural Networks [pdf] - Chunshui Cao, Xianming Liu, Yi Yang, Yinan Yu, Jiang Wang, Zilei Wang, Yongzhen Huang, Liang Wang, Chang Huang, Wei Xu, Deva Ramanan, Thomas S. Huang
  • Low-Rank Matrix Factorization Under General Mixture Noise Distributions [pdf] - Xiangyong Cao, Yang Chen, Qian Zhao, Deyu Meng, Yao Wang, Dong Wang, Zongben Xu
  • An Efficient Statistical Method for Image Noise Level Estimation [pdf] - Guangyong Chen, Fengyuan Zhu, Pheng Ann Heng
  • Webly Supervised Learning of Convolutional Networks [pdf] - Xinlei Chen, Abhinav Gupta
  • Love Thy Neighbors: Image Annotation by Exploiting Image Metadata [pdf] - Justin Johnson, Lamberto Ballan, Li Fei-Fei
Robust Image Segmentation Using Contour-Guided Color Palettes [pdf]
Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo

Abstract: The contour-guided color palette (CCP) 1 is proposed for robust image segmentation. It efficiently integrates contour and color cues of an image. To find representative colors of an image, color samples along long contours between regions, similar in spirit to machine learning methodology that focus on samples near decision boundaries, are collected followed by the mean-shift (MS) algorithm in the sampled color space to achieve an image-dependent color palette. This color palette provides a preliminary segmentation in the spatial domain, which is further fine-tuned by post-processing techniques such as leakage avoidance, fake boundary removal, and small region mergence. Segmentation performances of CCP and MS are compared and analyzed. While CCP offers an acceptable standalone segmentation result, it can be further integrated into the framework of layered spectral segmentation to produce a more robust segmentation. The superior performance of CCPbased segmentation algorithm is demonstrated by experiments on the Berkeley Segmentation Dataset.
Similar papers:
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
  • Illumination Robust Color Naming via Label Propagation [pdf] - Yuanliu liu, Zejian Yuan, Badong Chen, Jianru Xue, Nanning Zheng
  • Beyond White: Ground Truth Colors for Color Constancy Correction [pdf] - Dongliang Cheng, Brian Price, Scott Cohen, Michael S. Brown
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
Massively Parallel Multiview Stereopsis by Surface Normal Diffusion [pdf]
Silvano Galliani, Katrin Lasinger, Konrad Schindler

Abstract: We present a new, massively parallel method for highquality multiview matching. Our work builds on the Patchmatch idea: starting from randomly generated 3D planes in scene space, the best-fitting planes are iteratively propagated and refined to obtain a 3D depth and normal field per view, such that a robust photo-consistency measure over all images is maximized. Our main novelties are on the one hand to formulate Patchmatch in scene space, which makes it possible to aggregate image similarity across multiple views and obtain more accurate depth maps. And on the other hand a modified, diffusion-like propagation scheme that can be massively parallelized and delivers dense multiview correspondence over ten 1.9-Megapixel images in 3 seconds, on a consumer-grade GPU. Our method uses a slanted support window and thus has no fronto-parallel bias; it is completely local and parallel, such that computation time scales linearly with image size, and inversely proportional to the number of parallel threads. Furthermore, it has low memory footprint (four values per pixel, independent of the depth range). It therefore scales exceptionally well and can handle multiple large images at high depth resolution. Experiments on the DTU and Middlebury multiview datasets as well as oblique aerial images show that our method achieves very competitive results with high accuracy and completeness, across a range of different scenarios.
Similar papers:
  • MeshStereo: A Global Stereo Model With Mesh Alignment Regularization for View Interpolation [pdf] - Chi Zhang, Zhiwei Li, Yanhua Cheng, Rui Cai, Hongyang Chao, Yong Rui
  • A Deep Visual Correspondence Embedding Model for Stereo Matching Costs [pdf] - Zhuoyuan Chen, Xun Sun, Liang Wang, Yinan Yu, Chang Huang
  • SPM-BP: Sped-up PatchMatch Belief Propagation for Continuous MRFs [pdf] - Yu Li, Dongbo Min, Michael S. Brown, Minh N. Do, Jiangbo Lu
  • Exploiting High Level Scene Cues in Stereo Reconstruction [pdf] - Simon Hadfield, Richard Bowden
  • Variational PatchMatch MultiView Reconstruction and Refinement [pdf] - Philipp Heise, Brian Jensen, Sebastian Klose, Alois Knoll
Wide Baseline Stereo Matching With Convex Bounded Distortion Constraints [pdf]
Meirav Galun, Tal Amir, Tal Hassner, Ronen Basri, Yaron Lipman

Abstract: Finding correspondences in wide baseline setups is a challenging problem. Existing approaches have focused largely on developing better feature descriptors for correspondence and on accurate recovery of epipolar line constraints. This paper focuses on the challenging problem of finding correspondences once approximate epipolar constraints are given. We introduce a novel method that integrates a deformation model. Specifically, we formulate the problem as finding the largest number of corresponding points related by a bounded distortion map that obeys the given epipolar constraints. We show that, while the set of bounded distortion maps is not convex, the subset of maps that obey the epipolar line constraints is convex, allowing us to introduce an efficient algorithm for matching. We further utilize a robust cost function for matching and employ majorization-minimization for its optimization. Our experiments indicate that our method finds significantly more accurate maps than existing approaches.
Similar papers:
  • Efficient Solution to the Epipolar Geometry for Radially Distorted Cameras [pdf] - Zuzana Kukelova, Jan Heller, Martin Bujnak, Andrew Fitzgibbon, Tomas Pajdla
  • Exploiting High Level Scene Cues in Stereo Reconstruction [pdf] - Simon Hadfield, Richard Bowden
  • Minimal Solvers for 3D Geometry From Satellite Imagery [pdf] - Enliang Zheng, Ke Wang, Enrique Dunn, Jan-Michael Frahm
  • Multi-Image Matching via Fast Alternating Minimization [pdf] - Xiaowei Zhou, Menglong Zhu, Kostas Daniilidis
  • The Joint Image Handbook [pdf] - Matthew Trager, Martial Hebert, Jean Ponce
Multi-View Subspace Clustering [pdf]
Hongchang Gao, Feiping Nie, Xuelong Li, Heng Huang

Abstract: For many computer vision applications, the data sets distribute on certain low-dimensional subspaces. Subspace clustering is to find such underlying subspaces and cluster the data points correctly. In this paper, we propose a novel multi-view subspace clustering method. The proposed method performs clustering on the subspace representation of each view simultaneously. Meanwhile, we propose to use a common cluster structure to guarantee the consistence among different views. In addition, an efficient algorithm is proposed to solve the problem. Experiments on four benchmark data sets have been performed to validate our proposed method. The promising results demonstrate the effectiveness of our method.
Similar papers:
  • Discriminative Low-Rank Tracking [pdf] - Yao Sui, Yafei Tang, Li Zhang
  • Local Subspace Collaborative Tracking [pdf] - Lin Ma, Xiaoqin Zhang, Weiming Hu, Junliang Xing, Jiwen Lu, Jie Zhou
  • Shape Interaction Matrix Revisited and Robustified: Efficient Subspace Clustering With Corrupted and Incomplete Data [pdf] - Pan Ji, Mathieu Salzmann, Hongdong Li
  • Low-Rank Tensor Constrained Multiview Subspace Clustering [pdf] - Changqing Zhang, Huazhu Fu, Si Liu, Guangcan Liu, Xiaochun Cao
  • Temporal Subspace Clustering for Human Motion Segmentation [pdf] - Sheng Li, Kang Li, Yun Fu
Structured Feature Selection [pdf]
Tian Gao, Ziheng Wang, Qiang Ji

Abstract: Feature dimensionality reduction has been widely used in various computer vision tasks. We explore feature selection as the dimensionality reduction technique and propose to use a structured approach, based on the Markov Blanket (MB), to select features. We first introduce a new MB discovery algorithm, Simultaneous Markov Blanket (STMB) discovery, that improves the efficiency of state-of-the-art algorithms. Then we theoretically justify three advantages of structured feature selection over traditional feature selection methods. Specifically, we show that the Markov Blanket is the minimum feature set that retains the maximal mutual information and also gives the lowest Bayes classification error. Then we apply structured feature selection to two applications: 1) We introduce a new method that enables STMB to scale up and show the competitive performance of our algorithms on large-scale image classification tasks. 2) We propose a method for structured feature selection to handle hierarchical features and show the proposed method can lead to big performance gain in facial expression and action unit (AU) recognition tasks.
Similar papers:
  • From Emotions to Action Units With Hidden and Semi-Hidden-Task Learning [pdf] - Adria Ruiz, Joost Van de Weijer, Xavier Binefa
  • Joint Fine-Tuning in Deep Neural Networks for Facial Expression Recognition [pdf] - Heechul Jung, Sihaeng Lee, Junho Yim, Sunjeong Park, Junmo Kim
  • Mining And-Or Graphs for Graph Matching and Object Discovery [pdf] - Quanshi Zhang, Ying Nian Wu, Song-Chun Zhu
  • Visual Tracking With Fully Convolutional Networks [pdf] - Lijun Wang, Wanli Ouyang, Xiaogang Wang, Huchuan Lu
  • Infinite Feature Selection [pdf] - Giorgio Roffo, Simone Melzi, Marco Cristani
Person Re-Identification Ranking Optimisation by Discriminant Context Information Analysis [pdf]
Jorge Garcia, Niki Martinel, Christian Micheloni, Alfredo Gardel

Abstract: Person re-identification is an open and challenging problem in computer vision. Existing re-identification approaches focus on optimal methods for features matching (e.g., metric learning approaches) or study the inter-camera transformations of such features. These methods hardly ever pay attention to the problem of visual ambiguities shared between the first ranks. In this paper, we focus on such a problem and introduce an unsupervised ranking optimization approach based on discriminant context information analysis. The proposed approach refines a given initial ranking by removing the visual ambiguities common to first ranks. This is achieved by analyzing their content and context information. Extensive experiments on three publicly available benchmark datasets and different baseline methods have been conducted. Results demonstrate a remarkable improvement in the first positions of the ranking. Regardless of the selected dataset, state-of-the-art methods are strongly outperformed by our method.
Similar papers:
  • Scalable Person Re-Identification: A Benchmark [pdf] - Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian
  • Multi-Scale Learning for Low-Resolution Person Re-Identification [pdf] - Xiang Li, Wei-Shi Zheng, Xiaojuan Wang, Tao Xiang, Shaogang Gong
  • Multi-Task Learning With Low Rank Attribute Embedding for Person Re-Identification [pdf] - Chi Su, Fan Yang, Shiliang Zhang, Qi Tian, Larry S. Davis, Wen Gao
  • Partial Person Re-Identification [pdf] - Wei-Shi Zheng, Xiang Li, Tao Xiang, Shengcai Liao, Jianhuang Lai, Shaogang Gong
  • Person Re-Identification With Discriminatively Trained Viewpoint Invariant Dictionaries [pdf] - Srikrishna Karanam, Yang Li, Richard J. Radke
Active Transfer Learning With Zero-Shot Priors: Reusing Past Datasets for Future Tasks [pdf]
Efstratios Gavves, Thomas Mensink, Tatiana Tommasi, Cees G. M. Snoek, Tinne Tuytelaars

Abstract: How can we reuse existing knowledge, in the form of available datasets, when solving a new and apparently unrelated target task from a set of unlabeled data? In this work we make a first contribution to answer this question in the context of image classification. We frame this quest as an active learning problem and use zero-shot classifiers to guide the learning process by linking the new task to the existing classifiers. By revisiting the dual formulation of adaptive SVM, we reveal two basic conditions to choose greedily only the most relevant samples to be annotated. On this basis we propose an effective active learning algorithm which learns the best possible target classification model with minimum human labeling effort. Extensive experiments on two challenging datasets show the value of our approach compared to the state-of-the-art active learning methodologies, as well as its potential to reuse past datasets with minimal effort for future tasks.
Similar papers:
  • Efficient Classifier Training to Minimize False Merges in Electron Microscopy Segmentation [pdf] - Toufiq Parag, Dan C. Ciresan, Alessandro Giusti
  • Context Aware Active Learning of Activity Recognition Models [pdf] - Mahmudul Hasan, Amit K. Roy-Chowdhury
  • Point Triangulation Through Polyhedron Collapse Using the l[?] Norm [pdf] - Simon Donne, Bart Goossens, Wilfried Philips
  • Introducing Geometry in Active Learning for Image Segmentation [pdf] - Ksenia Konyushkova, Raphael Sznitman, Pascal Fua
  • Multi-Class Multi-Annotator Active Learning With Robust Gaussian Process for Visual Recognition [pdf] - Chengjiang Long, Gang Hua
Learning Image and User Features for Recommendation in Social Networks [pdf]
Xue Geng, Hanwang Zhang, Jingwen Bian, Tat-Seng Chua

Abstract: Sports Good representations of data do help in many machine learning tasks such as recommendation. It is often a great challenge for traditional recommender systems to learn representative features of both users and images in large social networks, in particular, social curation networks, which are characterized as the extremely sparse links between users and images, and the extremely diverse visual contents of images. To address the challenges, we propose a novel deep model which learns the unified feature representations for both users and images. This is done by transforming the heterogeneous user-image networks into homogeneous low-dimensional representations, which facilitate a recommender to trivially recommend images to users by feature similarity. We also develop a fast online algorithm that can be easily scaled up to large networks in an asynchronously parallel way. We conduct extensive experiments on a representative subset of Pinterest, containing 1,456,540 images and 1,000,000 users. Results of image recommendation experiments demonstrate that our feature learning approach significantly outperforms other state-of-the-art recommendation methods.
Similar papers:
  • Learning Visual Clothing Style With Heterogeneous Dyadic Co-Occurrences [pdf] - Andreas Veit, Balazs Kovacs, Sean Bell, Julian McAuley, Kavita Bala, Serge Belongie
  • Panoptic Studio: A Massively Multiview System for Social Motion Capture [pdf] - Hanbyul Joo, Hao Liu, Lei Tan, Lin Gui, Bart Nabbe, Iain Matthews, Takeo Kanade, Shohei Nobuhara, Yaser Sheikh
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
  • Learning Social Relation Traits From Face Images [pdf] - Zhanpeng Zhang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Automated Facial Trait Judgment and Election Outcome Prediction: Social Dimensions of Face [pdf] - Jungseock Joo, Francis F. Steen, Song-Chun Zhu
A Gaussian Process Latent Variable Model for BRDF Inference [pdf]
Stamatios Georgoulis, Vincent Vanweddingen, Marc Proesmans, Luc Van Gool

Abstract: The problem of estimating a full BRDF from partial observations has already been studied using either parametric or non-parametric approaches. The goal in each case is to best match this sparse set of input measurements. In this paper we address the problem of inferring higher order reflectance information starting from the minimal input of a single BRDF slice. We begin from the prototypical case of a homogeneous sphere, lit by a head-on light source, which only holds information about less than 0.001% of the whole BRDF domain. We propose a novel method to infer the higher dimensional properties of the materials BRDF, based on the statistical distribution of known material characteristics observed in real-life samples. We evaluated our method based on a large set of experiments generated from real-world BRDFs and newly measured materials. Although inferring higher dimensional BRDFs from such modest training is not a trivial problem, our method performs better than state-of-the-art parametric, semi- parametric and non-parametric approaches. Finally, we discuss interesting applications on material relighting, and flash-based photography.
Similar papers:
  • A Comprehensive Multi-Illuminant Dataset for Benchmarking of the Intrinsic Image Algorithms [pdf] - Shida Beigpour, Andreas Kolb, Sven Kunz
  • Separating Fluorescent and Reflective Components by Using a Single Hyperspectral Image [pdf] - Yinqiang Zheng, Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • Learning Ordinal Relationships for Mid-Level Vision [pdf] - Daniel Zoran, Phillip Isola, Dilip Krishnan, William T. Freeman
  • Intrinsic Decomposition of Image Sequences From Local Temporal Variations [pdf] - Pierre-Yves Laffont, Jean-Charles Bazin
  • Learning Data-Driven Reflectance Priors for Intrinsic Image Decomposition [pdf] - Tinghui Zhou, Philipp Krahenbuhl, Alexei A. Efros
Domain Generalization for Object Recognition With Multi-Task Autoencoders [pdf]
Muhammad Ghifary, W. Bastiaan Kleijn, Mengjie Zhang, David Balduzzi

Abstract: The problem of domain generalization is to take knowledge acquired from a number of related domains where training data is available, and to then successfully apply it to previously unseen domains. We propose a new feature learning algorithm, Multi-Task Autoencoder (MTAE), that provides good generalization performance for crossdomain object recognition. Our algorithm extends the standard denoising autoencoder framework by substituting artificially induced corruption with naturally occurring inter-domain variability in the appearance of objects. Instead of reconstructing images from noisy versions, MTAE learns to transform the original image into analogs in multiple related domains. It thereby learns features that are robust to variations across domains. The learnt features are then used as inputs to a classifier. We evaluated the performance of the algorithm on benchmark image recognition datasets, where the task is to learn features from multiple datasets and to then predict the image label from unseen datasets. We found that (denoising) MTAE outperforms alternative autoencoder-based models as well as the current state-of-the-art algorithms for domain generalization.
Similar papers:
  • Zero-Shot Learning via Semantic Similarity Embedding [pdf] - Ziming Zhang, Venkatesh Saligrama
  • Simultaneous Deep Transfer Across Domains and Tasks [pdf] - Eric Tzeng, Judy Hoffman, Trevor Darrell, Kate Saenko
  • Unsupervised Domain Adaptation With Imbalanced Cross-Domain Data [pdf] - Tzu Ming Harry Hsu, Wei Yu Chen, Cheng-An Hou, Yao-Hung Hubert Tsai, Yi-Ren Yeh, Yu-Chiang Frank Wang
  • Bi-Shifting Auto-Encoder for Unsupervised Domain Adaptation [pdf] - Meina Kan, Shiguang Shan, Xilin Chen
  • Multi-View Domain Generalization for Visual Recognition [pdf] - Li Niu, Wen Li, Dong Xu
DeepProposal: Hunting Objects by Cascading Deep Convolutional Layers [pdf]
Amir Ghodrati, Ali Diba, Marco Pedersoli, Tinne Tuytelaars, Luc Van Gool

Abstract: In this paper we evaluate the quality of the activation layers of a convolutional neural network (CNN) for the generation of object proposals. We generate hypotheses in a sliding-window fashion over different activation layers and show that the final convolutional layers can find the object of interest with high recall but poor localization due to the coarseness of the feature maps. Instead, the first layers of the network can better localize the object of interest but with a reduced recall. Based on this observation we design a method for proposing object locations that is based on CNN features and that combines the best of both worlds. We build an inverse cascade that, going from the final to the initial convolutional layers of the CNN, selects the most promising object locations and refines their boxes in a coarse-to-fine manner. The method is efficient, because i) it uses the same features extracted for detection, ii) it aggregates features using integral images, and iii) it avoids a dense evaluation of the proposals due to the inverse coarse-to-fine cascade. The method is also accurate; it outperforms most of the previously proposed object proposals approaches and when plugged into a CNN-based detector produces state-of-theart detection performance.
Similar papers:
  • Contour Box: Rejecting Object Proposals Without Explicit Closed Contours [pdf] - Cewu Lu, Shu Liu, Jiaya Jia, Chi-Keung Tang
  • Box Aggregation for Proposal Decimation: Last Mile of Object Detection [pdf] - Shu Liu, Cewu Lu, Jiaya Jia
  • Oriented Object Proposals [pdf] - Shengfeng He, Rynson W.H. Lau
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • Cascaded Sparse Spatial Bins for Efficient and Effective Generic Object Detection [pdf] - David Novotny, Jiri Matas
Object Detection via a Multi-Region and Semantic Segmentation-Aware CNN Model [pdf]
Spyros Gidaris, Nikos Komodakis

Abstract: We propose an object detection system that relies on a multi-region deep convolutional neural network (CNN) that also encodes semantic segmentation-aware features. The resulting CNN-based representation aims at capturing a diverse set of discriminative appearance factors and exhibits localization sensitivity that is essential for accurate object localization. We exploit the above properties of our recognition module by integrating it on an iterative localization mechanism that alternates between scoring a box proposal and refining its location with a deep CNN regression model. Thanks to the efficient use of our modules, we detect objects with very high localization accuracy. On the detection challenges of PASCAL VOC2007 and PASCAL VOC2012 we achieve mAP of 78.2% and 73.9% correspondingly, surpassing any other published work by a significant margin.
Similar papers:
  • Contextual Action Recognition With R*CNN [pdf] - Georgia Gkioxari, Ross Girshick, Jitendra Malik
  • Multiple Granularity Descriptors for Fine-Grained Categorization [pdf] - Dequan Wang, Zhiqiang Shen, Jie Shao, Wei Zhang, Xiangyang Xue, Zheng Zhang
  • Active Object Localization With Deep Reinforcement Learning [pdf] - Juan C. Caicedo, Svetlana Lazebnik
  • AttentionNet: Aggregating Weak Directions for Accurate Object Detection [pdf] - Donggeun Yoo, Sunggyun Park, Joon-Young Lee, Anthony S. Paek, In So Kweon
  • BoxSup: Exploiting Bounding Boxes to Supervise Convolutional Networks for Semantic Segmentation [pdf] - Jifeng Dai, Kaiming He, Jian Sun
Fast R-CNN [pdf]
Ross Girshick

Abstract: This paper proposes a Fast Region-based Convolutional Network method (Fast R-CNN) for object detection. Fast R-CNN builds on previous work to efficiently classify object proposals using deep convolutional networks. Compared to previous work, Fast R-CNN employs several innovations to improve training and testing speed while also increasing detection accuracy. Fast R-CNN trains the very deep VGG16 network 9 faster than R-CNN, is 213 faster at test-time, and achieves a higher mAP on PASCAL VOC 2012. Compared to SPPnet, Fast R-CNN trains VGG16 3 faster, tests 10 faster, and is more accurate. Fast R-CNN is implemented in Python and C++ (using Caffe) and is available under the open-source MIT License at https: //github.com/rbgirshick/fast-rcnn. while achieving top accuracy on PASCAL VOC 2012 [7] with a mAP of 66% (vs. 62% for R-CNN).1 1.1. R-CNN and SPPnet The Region-based Convolutional Network method (RCNN) [9] achieves excellent object detection accuracy by using a deep ConvNet to classify object proposals. R-CNN, however, has notable drawbacks: 1. Training is a multi-stage pipeline. R-CNN first finetunes a ConvNet on object proposals using log loss. Then, it fits SVMs to ConvNet features. These SVMs act as object detectors, replacing the softmax classifier learnt by fine-tuning. In the third training stage, bounding-box regressors are learned. 2. Training is expensive in space and time. For SVM and bounding-box regressor training, features are extracted from each object pr
Similar papers:
  • Multi-Scale Recognition With DAG-CNNs [pdf] - Songfan Yang, Deva Ramanan
  • Deep Fried Convnets [pdf] - Zichao Yang, Marcin Moczulski, Misha Denil, Nando de Freitas, Alex Smola, Le Song, Ziyu Wang
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • DeepProposal: Hunting Objects by Cascading Deep Convolutional Layers [pdf] - Amir Ghodrati, Ali Diba, Marco Pedersoli, Tinne Tuytelaars, Luc Van Gool
  • An Exploration of Parameter Redundancy in Deep Networks With Circulant Projections [pdf] - Yu Cheng, Felix X. Yu, Rogerio S. Feris, Sanjiv Kumar, Alok Choudhary, Shi-Fu Chang
Actions and Attributes From Wholes and Parts [pdf]
Georgia Gkioxari, Ross Girshick, Jitendra Malik

Abstract: We investigate the importance of parts for the tasks of action and attribute classification. We develop a part-based approach by leveraging convolutional network features inspired by recent advances in computer vision. Our part detectors are a deep version of poselets and capture parts of the human body under a distinct set of poses. For the tasks of action and attribute classification, we train holistic convolutional neural networks and show that adding parts leads to top-performing results for both tasks. We observe that for deeper networks parts are less significant. In addition, we demonstrate the effectiveness of our approach when we replace an oracle person detector, as is the default in the current evaluation protocol for both tasks, with a state-ofthe-art person detection system.
Similar papers:
  • Learning Deep Representation With Large-Scale Attributes [pdf] - Wanli Ouyang, Hongyang Li, Xingyu Zeng, Xiaogang Wang
  • An MRF-Poselets Model for Detecting Highly Articulated Humans [pdf] - Duc Thanh Nguyen, Minh-Khoi Tran, Sai-Kit Yeung
  • Neural Activation Constellations: Unsupervised Part Model Discovery With Convolutional Networks [pdf] - Marcel Simon, Erik Rodner
  • Attributed Grammars for Joint Estimation of Human Attributes, Part and Pose [pdf] - Seyoung Park, Song-Chun Zhu
  • Contextual Action Recognition With R*CNN [pdf] - Georgia Gkioxari, Ross Girshick, Jitendra Malik
Contextual Action Recognition With R*CNN [pdf]
Georgia Gkioxari, Ross Girshick, Jitendra Malik

Abstract: There are multiple cues in an image which reveal what action a person is performing. For example, a jogger has a pose that is characteristic for jogging, but the scene (e.g. road, trail) and the presence of other joggers can be an additional source of information. In this work, we exploit the simple observation that actions are accompanied by contextual cues to build a strong action recognition system. We adapt RCNN to use more than one region for classification while still maintaining the ability to localize the action. We call our system R CNN. The action-specific models and the feature maps are trained jointly, allowing for action specific representations to emerge. R CNN achieves 90.2% mean AP on the PASAL VOC Action dataset, outperforming all other approaches in the field by a significant margin. Last, we show that R CNN is not limited to action recognition. In particular, R CNN can also be used to tackle fine-grained tasks such as attribute classification. We validate this claim by reporting state-of-the-art performance on the Berkeley Attributes of People dataset.1
Similar papers:
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
  • Actions and Attributes From Wholes and Parts [pdf] - Georgia Gkioxari, Ross Girshick, Jitendra Malik
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
Hot or Not: Exploring Correlations Between Appearance and Temperature [pdf]
Daniel Glasner, Pascal Fua, Todd Zickler, Lihi Zelnik-Manor

Abstract: In this paper we explore interactions between the appearance of an outdoor scene and the ambient temperature. By studying statistical correlations between image sequences from outdoor cameras and temperature measurements we identify two interesting interactions. First, semantically meaningful regions such as foliage and reflective oriented surfaces are often highly indicative of the temperature. Second, small camera motions are correlated with the temperature in some scenes. We propose simple scenespecific temperature prediction algorithms which can be used to turn a camera into a crude temperature sensor. We find that for this task, simple features such as local pixel intensities outperform sophisticated, global features such as from a semantically-trained convolutional neural network. (a) Claude Monet, Grainstacks at (b) Claude Monet, Grainstacks, the End of Summer, Morning Effect Snow Effect (c) hot... (d) or not... Figure 1: How do we perceive temperature differences?
Similar papers:
  • 3D Time-Lapse Reconstruction From Internet Photos [pdf] - Ricardo Martin-Brualla, David Gallup, Steven M. Seitz
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
  • Building Dynamic Cloud Maps From the Ground Up [pdf] - Calvin Murdock, Nathan Jacobs, Robert Pless
  • PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization [pdf] - Alex Kendall, Matthew Grimes, Roberto Cipolla
  • Synthesizing Illumination Mosaics From Internet Photo-Collections [pdf] - Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm
LEWIS: Latent Embeddings for Word Images and their Semantics [pdf]
Albert Gordo, Jon Almazan, Naila Murray, Florent Perronin

Abstract: The goal of this work is to bring semantics into the tasks of text recognition and retrieval in natural images. Although text recognition and retrieval have received a lot of attention in recent years, previous works have focused on recognizing or retrieving exactly the same word used as a query, without taking the semantics into consideration. In this paper, we ask the following question: can we predict semantic concepts directly from a word image, without explicitly trying to transcribe the word image or its characters at any point? For this goal we propose a convolutional neural network (CNN) with a weighted ranking loss objective that ensures that the concepts relevant to the query image are ranked ahead of those that are not relevant. This can also be interpreted as learning a Euclidean space where word images and concepts are jointly embedded. This model is learned in an end-to-end manner, from image pixels to semantic concepts, using a dataset of synthetically generated word images and concepts mined from a lexical database (WordNet). Our results show that, despite the complexity of the task, word images and concepts can indeed be associated with a high degree of accuracy.
Similar papers:
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Predicting Deep Zero-Shot Convolutional Neural Networks Using Textual Descriptions [pdf] - Jimmy Lei Ba, Kevin Swersky, Sanja Fidler, Ruslan salakhutdinov
  • Selecting Relevant Web Trained Concepts for Automated Event Retrieval [pdf] - Bharat Singh, Xintong Han, Zhe Wu, Vlad I. Morariu, Larry S. Davis
  • Automatic Concept Discovery From Parallel Text and Visual Corpora [pdf] - Chen Sun, Chuang Gan, Ram Nevatia
  • Learning Like a Child: Fast Novel Visual Concept Learning From Sentence Descriptions of Images [pdf] - Junhua Mao, Xu Wei, Yi Yang, Jiang Wang, Zhiheng Huang, Alan L. Yuille
Unsupervised Learning of Spatiotemporally Coherent Metrics [pdf]
Ross Goroshin, Joan Bruna, Jonathan Tompson, David Eigen, Yann LeCun

Abstract: Current state-of-the-art classification and detection algorithms train deep convolutional networks using labeled data. In this work we study unsupervised feature learning with convolutional networks in the context of temporally coherent unlabeled data. We focus on feature learning from unlabeled video data, using the assumption that adjacent video frames contain semantically similar information. This assumption is exploited to train a convolutional pooling auto-encoder regularized by slowness and sparsity priors. We establish a connection between slow feature learning and metric learning. Using this connection we define temporal coherencea criterion which can be used to set hyper-parameters in a principled and automated manner. In a transfer learning experiment, we show that the resulting encoder can be used to define a more semantically coherent metric without the use of labels.
Similar papers:
  • Learning Spatiotemporal Features With 3D Convolutional Networks [pdf] - Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, Manohar Paluri
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
  • Task-Driven Feature Pooling for Image Classification [pdf] - Guo-Sen Xie, Xu-Yao Zhang, Xiangbo Shu, Shuicheng Yan, Cheng-Lin Liu
  • Learning Image Representations Tied to Ego-Motion [pdf] - Dinesh Jayaraman, Kristen Grauman
Photogeometric Scene Flow for High-Detail Dynamic 3D Reconstruction [pdf]
Paulo F. U. Gotardo, Tomas Simon, Yaser Sheikh, Iain Matthews

Abstract: Photometric stereo (PS) is an established technique for high-detail reconstruction of 3D geometry and appearance. To correct for surface integration errors, PS is often combined with multiview stereo (MVS). With dynamic objects, PS reconstruction also faces the problem of computing optical flow (OF) for image alignment under rapid changes in illumination. Current PS methods typically compute optical flow and MVS as independent stages, each one with its own limitations and errors introduced by early regularization. In contrast, scene flow methods estimate geometry and motion, but lack the fine detail from PS. This paper proposes photogeometric scene flow (PGSF) for high-quality dynamic 3D reconstruction. PGSF performs PS, OF, and MVS simultaneously. It is based on two key observations: (i) while image alignment improves PS, PS allows for surfaces to be relit to improve alignment; (ii) PS provides surface gradients that render the smoothness term in MVS unnecessary, leading to truly data-driven, continuous depth estimates. This synergy is demonstrated in the quality of the resulting RGB appearance, 3D geometry, and 3D motion.
Similar papers:
  • Flow Fields: Dense Correspondence Fields for Highly Accurate Large Displacement Optical Flow Estimation [pdf] - Christian Bailer, Bertram Taetz, Didier Stricker
  • Large Displacement 3D Scene Flow With Occlusion Reasoning [pdf] - Andrei Zanfir, Cristian Sminchisescu
  • Photometric Stereo in a Scattering Medium [pdf] - Zak Murez, Tali Treibitz, Ravi Ramamoorthi, David Kriegman
  • Photometric Stereo With Small Angular Variations [pdf] - Jian Wang, Yasuyuki Matsushita, Boxin Shi, Aswin C. Sankaranarayanan
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
Convolutional Sparse Coding for Image Super-Resolution [pdf]
Shuhang Gu, Wangmeng Zuo, Qi Xie, Deyu Meng, Xiangchu Feng, Lei Zhang

Abstract: Most of the previous sparse coding (SC) based super resolution (SR) methods partition the image into overlapped patches, and process each patch separately. These methods, however, ignore the consistency of pixels in overlapped patches, which is a strong constraint for image reconstruction. In this paper, we propose a convolutional sparse coding (CSC) based SR (CSC-SR) method to address the consistency issue. Our CSC-SR involves three groups of parameters to be learned: (i) a set of filters to decompose the low resolution (LR) image into LR sparse feature maps; (ii) a mapping function to predict the high resolution (HR) feature maps from the LR ones; and (iii) a set of filters to reconstruct the HR images from the predicted HR feature maps via simple convolution operations. By working directly on the whole image, the proposed CSC-SR algorithm does not need to divide the image into overlapped patches, and can exploit the image global correlation to produce more robust reconstruction of image local structures. Experimental results clearly validate the advantages of CSC over patch based SC in SR application. Compared with state-of-the-art SR methods, the proposed CSC-SR method achieves highly competitive PSNR results, while demonstrating better edge and texture preservation performance.
Similar papers:
  • Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf] - Yuhui Quan, Yan Huang, Hui Ji
  • Learning Parametric Distributions for Image Super-Resolution: Where Patch Matching Meets Sparse Coding [pdf] - Yongbo Li, Weisheng Dong, Guangming Shi, Xuemei Xie
  • Variational Depth Superresolution Using Example-Based Edge Representations [pdf] - David Ferstl, Matthias Ruther, Horst Bischof
  • Conformal and Low-Rank Sparse Representation for Image Restoration [pdf] - Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng
  • Deep Networks for Image Super-Resolution With Sparse Prior [pdf] - Zhaowen Wang, Ding Liu, Jianchao Yang, Wei Han, Thomas Huang
Learning to Boost Filamentary Structure Segmentation [pdf]
Lin Gu, Li Cheng

Abstract: The challenging problem of filamentary structure segmentation has a broad range of applications in biological and medical fields. A critical yet challenging issue remains on how to detect and restore the small filamentary fragments from backgrounds: The small fragments are of diverse shapes and appearances, meanwhile the backgrounds could be cluttered and ambiguous. Focusing on this issue, this paper proposes an iterative two-step learning-based approach to boost the performance based on a base segmenter arbitrarily chosen from a number of existing segmenters: We start with an initial partial segmentation where the filamentary structure obtained is of high confidence based on this existing segmenter. We also define a scanning horizon as epsilon balls centred around the partial segmentation result. Step one of our approach centers on a data-driven latent classification tree model to detect the filamentary fragments. This model is learned via a training process, where a large number of distinct local figure/background separation scenarios are established and geometrically organized into a tree structure. Step two spatially restores the isolated fragments back to the current partial segmentation, which is accomplished by means of completion fields and matting. Both steps are then alternated with the growth of partial segmentation result, until the input image space is entirely explored. Our approach is rather generic and can be easily augmented to a wide range of existing superv
Similar papers:
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
  • Efficient Classifier Training to Minimize False Merges in Electron Microscopy Segmentation [pdf] - Toufiq Parag, Dan C. Ciresan, Alessandro Giusti
  • Image Matting With KL-Divergence Based Sparse Sampling [pdf] - Levent Karacan, Aykut Erdem, Erkut Erdem
  • Contour Guided Hierarchical Model for Shape Matching [pdf] - Yuanqi Su, Yuehu Liu, Bonan Cuan, Nanning Zheng
  • 3D Fragment Reassembly Using Integrated Template Guidance and Fracture-Region Matching [pdf] - Kang Zhang, Wuyi Yu, Mary Manhein, Warren Waggenspack, Xin Li
Robust Non-Rigid Motion Tracking and Surface Reconstruction Using L0 Regularization [pdf]
Kaiwen Guo, Feng Xu, Yangang Wang, Yebin Liu, Qionghai Dai

Abstract: We present a new motion tracking method to robustly reconstruct non-rigid geometries and motions from single view depth inputs captured by a consumer depth sensor. The idea comes from the observation of the existence of intrinsic articulated subspace in most of non-rigid motions. To take advantage of this characteristic, we propose a novel L0 based motion regularizer with an iterative optimization solver that can implicitly constrain local deformation only on joints with articulated motions, leading to reduced solution space and physical plausible deformations. The L0 strategy is integrated into the available non-rigid motion tracking pipeline, forming the proposed L0 -L2 non-rigid motion tracking method that can adaptively stop the tracking error propagation. Extensive experiments over complex human body motions with occlusions, face and hand motions demonstrate that our approach substantially improves tracking robustness and surface reconstruction accuracy.
Similar papers:
  • A Versatile Learning-Based 3D Temporal Tracker: Scalable, Robust, Online [pdf] - David Joseph Tan, Federico Tombari, Slobodan Ilic, Nassir Navab
  • Self-Occlusions and Disocclusions in Causal Video Object Segmentation [pdf] - Yanchao Yang, Ganesh Sundaramoorthi, Stefano Soatto
  • Integrating Dashcam Views Through Inter-Video Mapping [pdf] - Hsin-I Chen, Yi-Ling Chen, Wei-Tse Lee, Fan Wang, Bing-Yu Chen
  • Recurrent Network Models for Human Dynamics [pdf] - Katerina Fragkiadaki, Sergey Levine, Panna Felsen, Jitendra Malik
  • Model-Based Tracking at 300Hz Using Raw Time-of-Flight Observations [pdf] - Jan Stuhmer, Sebastian Nowozin, Andrew Fitzgibbon, Richard Szeliski, Travis Perry, Sunil Acharya, Daniel Cremers, Jamie Shotton
Accurate Camera Calibration Robust to Defocus Using a Smartphone [pdf]
Hyowon Ha, Yunsu Bok, Kyungdon Joo, Jiyoung Jung, In So Kweon

Abstract: We propose a novel camera calibration method for defocused images using a smartphone under the assumption that the defocus blur is modeled as a convolution of a sharp image with a Gaussian point spread function (PSF). In contrast to existing calibration approaches which require wellfocused images, the proposed method achieves accurate camera calibration with severely defocused images. This robustness to defocus is due to the proposed set of unidirectional binary patterns, which simplifies 2D Gaussian deconvolution to a 1D Gaussian deconvolution problem with multiple observations. By capturing the set of patterns consecutively displayed on a smartphone, we formulate the feature extraction as a deconvolution problem to estimate feature point locations in sub-pixel accuracy and the blur kernel in each location. We also compensate the error in camera parameters due to refraction of the glass panel of the display device. We evaluate the performance of the proposed method on synthetic and real data. Even under severe defocus, our method shows accurate camera calibration result.
Similar papers:
  • Blur-Aware Disparity Estimation From Defocus Stereo Images [pdf] - Ching-Hui Chen, Hui Zhou, Timo Ahonen
  • SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf] - Kensei Jo, Mohit Gupta, Shree K. Nayar
  • On the Equivalence of Moving Entrance Pupil and Radial Distortion for Camera Calibration [pdf] - Avinash Kumar, Narendra Ahuja
  • Non-Parametric Structure-Based Calibration of Radially Symmetric Cameras [pdf] - Federico Camposeco, Torsten Sattler, Marc Pollefeys
  • A Linear Generalized Camera Calibration From Three Intersecting Reference Planes [pdf] - Mai Nishimura, Shohei Nobuhara, Takashi Matsuyama, Shinya Shimizu, Kensaku Fujii
Intrinsic Scene Decomposition From RGB-D images [pdf]
Mohammed Hachama, Bernard Ghanem, Peter Wonka

Abstract: In this paper, we address the problem of computing an intrinsic decomposition of the colors of a surface into an albedo and a shading term. The surface is reconstructed from a single or multiple RGB-D images of a static scene obtained from different views. We thereby extend and improve existing works in the area of intrinsic image decomposition. In a variational framework, we formulate the problem as a minimization of an energy composed of two terms: a data term and a regularity term. The first term is related to the image formation process and expresses the relation between the albedo, the surface normals, and the incident illumination. We use an affine shading model, a combination of a Lambertian model, and an ambient lighting term. This model is relevant for Lambertian surfaces. When available, multiple views can be used to handle view-dependent nonLambertian reflections. The second term contains an efficient combination of 2 and 1 -regularizers on the illumination vector field and albedo respectively. Unlike most previous approaches, especially Retinex-like techniques, these terms do not depend on the image gradient or texture, thus reducing the mixing shading/reflectance artifacts and leading to better results. The obtained non-linear optimization problem is efficiently solved using a cyclic block coordinate descent algorithm. Our method outperforms a range of state-of-the-art algorithms on a popular benchmark dataset.
Similar papers:
  • Learning Data-Driven Reflectance Priors for Intrinsic Image Decomposition [pdf] - Tinghui Zhou, Philipp Krahenbuhl, Alexei A. Efros
  • Intrinsic Decomposition of Image Sequences From Local Temporal Variations [pdf] - Pierre-Yves Laffont, Jean-Charles Bazin
  • Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf] - Naejin Kong, Michael J. Black
  • A Comprehensive Multi-Illuminant Dataset for Benchmarking of the Intrinsic Image Algorithms [pdf] - Shida Beigpour, Andreas Kolb, Sven Kunz
  • Direct Intrinsics: Learning Albedo-Shading Decomposition by Convolutional Regression [pdf] - Takuya Narihira, Michael Maire, Stella X. Yu
Exploiting High Level Scene Cues in Stereo Reconstruction [pdf]
Simon Hadfield, Richard Bowden

Abstract: We present a novel approach to 3D reconstruction which is inspired by the human visual system. This system unifies standard appearance matching and triangulation techniques with higher level reasoning and scene understanding, in order to resolve ambiguities between different interpretations of the scene. The types of reasoning integrated in the approach includes recognising common configurations of surface normals and semantic edges (e.g. convex, concave and occlusion boundaries). We also recognise the coplanar, collinear and symmetric structures which are especially common in man made environments. (a) Input data (b) Appearance matching (c) Scene reasoning (d) Resulting reconstruction
Similar papers:
  • General Dynamic Scene Reconstruction From Multiple View Video [pdf] - Armin Mustafa, Hansung Kim, Jean-Yves Guillemaut, Adrian Hilton
  • Depth Map Estimation and Colorization of Anaglyph Images Using Local Color Prior and Reverse Intensity Distribution [pdf] - Williem, Ramesh Raskar, In Kyu Park
  • Variational PatchMatch MultiView Reconstruction and Refinement [pdf] - Philipp Heise, Brian Jensen, Sebastian Klose, Alois Knoll
  • A Deep Visual Correspondence Embedding Model for Stereo Matching Costs [pdf] - Zhuoyuan Chen, Xun Sun, Liang Wang, Yinan Yu, Chang Huang
  • Massively Parallel Multiview Stereopsis by Surface Normal Diffusion [pdf] - Silvano Galliani, Katrin Lasinger, Konrad Schindler
Learning Ensembles of Potential Functions for Structured Prediction With Latent Variables [pdf]
Hossein Hajimirsadeghi, Greg Mori

Abstract: Activity Label Hidden Variables Many visual recognition tasks involve modeling variables which are structurally related. Hidden conditional random fields (HCRFs) are a powerful class of models for encoding structure in weakly supervised training examples. This paper presents HCRF-Boost, a novel and general framework for learning HCRFs in functional space. An algorithm is proposed to learn the potential functions of an HCRF as a combination of abstract nonlinear feature functions, expressed by regression models. Consequently, the resulting latent structured model is not restricted to traditional log-linear potential functions or any explicit parameterization. Further, functional optimization helps to avoid direct interactions with the possibly large parameter space of nonlinear models and improves efficiency. As a result, a complex and flexible ensemble method is achieved for structured prediction which can be successfully used in a variety of applications. We validate the effectiveness of this method on tasks such as group activity recognition, human action recognition, and multi-instance learning of video events.
Similar papers:
  • Unsupervised Semantic Parsing of Video Collections [pdf] - Ozan Sener, Amir R. Zamir, Silvio Savarese, Ashutosh Saxena
  • Activity Auto-Completion: Predicting Human Activities From Partial Videos [pdf] - Zhen Xu, Laiyun Qing, Jun Miao
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
  • Entropy-Based Latent Structured Output Prediction [pdf] - Diane Bouchacourt, Sebastian Nowozin, M. Pawan Kumar
Beyond Gauss: Image-Set Matching on the Riemannian Manifold of PDFs [pdf]
Mehrtash Harandi, Mathieu Salzmann, Mahsa Baktashmotlagh

Abstract: State-of-the-art image-set matching techniques typically implicitly model each image-set with a Gaussian distribution. Here, we propose to go beyond these representations and model image-sets as probability distribution functions (PDFs) using kernel density estimators. To compare and match image-sets, we exploit Csiszar f -divergences, which bear strong connections to the geodesic distance defined on the space of PDFs, i.e., the statistical manifold. Furthermore, we introduce valid positive definite kernels on the statistical manifold, which let us make use of more powerful classification schemes to match image-sets. Finally, we introduce a supervised dimensionality reduction technique that learns a latent space where f -divergences reflect the class labels of the data. Our experiments on diverse problems, such as video-based face recognition and dynamic texture classification, evidence the benefits of our approach over the state-of-the-art image-set matching methods.
Similar papers:
  • Scalable Nonlinear Embeddings for Semantic Category-Based Image Retrieval [pdf] - Gaurav Sharma, Bernt Schiele
  • Shell PCA: Statistical Shape Modelling in Shell Space [pdf] - Chao Zhang, Behrend Heeren, Martin Rumpf, William A. P. Smith
  • Low Dimensional Explicit Feature Maps [pdf] - Ondrej Chum
  • Beyond Covariance: Feature Representation With Nonlinear Kernel Matrices [pdf] - Lei Wang, Jianjia Zhang, Luping Zhou, Chang Tang, Wanqing Li
  • Recursive Frechet Mean Computation on the Grassmannian and its Applications to Computer Vision [pdf] - Rudrasis Chakraborty, Baba C. Vemuri
Context Aware Active Learning of Activity Recognition Models [pdf]
Mahmudul Hasan, Amit K. Roy-Chowdhury

Abstract: Activity recognition in video has recently benefited from the use of the context e.g., inter-relationships among the activities and objects. However, these approaches require data to be labeled and entirely available at the outset. In contrast, we formulate a continuous learning framework for context aware activity recognition from unlabeled video data which has two distinct advantages over most existing methods. First, we propose a novel active learning technique which not only exploits the informativeness of the individual activity instances but also utilizes their contextual information during the query selection process; this leads to significant reduction in expensive manual annotation effort. Second, the learned models can be adapted online as more data is available. We formulate a conditional random field (CRF) model that encodes the context and devise an information theoretic approach that utilizes entropy and mutual information of the nodes to compute the set of most informative query instances, which need to be labeled by a human. These labels are combined with graphical inference techniques for incrementally updating the model as new videos come in. Experiments on four challenging datasets demonstrate that our framework achieves superior performance with significantly less amount of manual labeling.
Similar papers:
  • Active Transfer Learning With Zero-Shot Priors: Reusing Past Datasets for Future Tasks [pdf] - Efstratios Gavves, Thomas Mensink, Tatiana Tommasi, Cees G. M. Snoek, Tinne Tuytelaars
  • Multi-Class Multi-Annotator Active Learning With Robust Gaussian Process for Visual Recognition [pdf] - Chengjiang Long, Gang Hua
  • Temporal Perception and Prediction in Ego-Centric Video [pdf] - Yipin Zhou, Tamara L. Berg
  • Unsupervised Semantic Parsing of Video Collections [pdf] - Ozan Sener, Amir R. Zamir, Silvio Savarese, Ashutosh Saxena
  • Activity Auto-Completion: Predicting Human Activities From Partial Videos [pdf] - Zhen Xu, Laiyun Qing, Jun Miao
Multiple-Hypothesis Affine Region Estimation With Anisotropic LoG Filters [pdf]
Takahiro Hasegawa, Mitsuru Ambai, Kohta Ishikawa, Gou Koutaki, Yuji Yamauchi, Takayoshi Yamashita, Hironobu Fujiyoshi

Abstract: We propose a method for estimating multiple-hypothesis affine regions from a keypoint by using an anisotropic Laplacian-of-Gaussian (LoG) filter. Although conventional affine region detectors, such as Hessian/Harris-Affine, iterate to find an affine region that fits a given image patch, such iterative searching is adversely affected by an initial point. To avoid this problem, we allow multiple detections from a single keypoint. We demonstrate that the responses of all possible anisotropic LoG filters can be efficiently computed by factorizing them in a similar manner to spectral SIFT. A large number of LoG filters that are densely sampled in a parameter space are reconstructed by a weighted combination of a limited number of representative filters, called eigenfilters, by using singular value decomposition. Also, the reconstructed filter responses of the sampled parameters can be interpolated to a continuous representation by using a series of proper functions. This results in efficient multiple extrema searching in a continuous space. Experiments revealed that our method has higher repeatability than the conventional methods. Figure 1. Comparison of multiple affine region estimation results. Left column: detection from binary image. Central column: keypoint localization offsets. Right column: detection from grayscale image. space constructed by Gaussian kernels of different sizes. The Hessian-Laplace and Harris-Laplace keypoint detectors [14] also estimate scale using th
Similar papers:
  • Fully Connected Guided Image Filtering [pdf] - Longquan Dai, Mengke Yuan, Feihu Zhang, Xiaopeng Zhang
  • Learning a Descriptor-Specific 3D Keypoint Detector [pdf] - Samuele Salti, Federico Tombari, Riccardo Spezialetti, Luigi Di Stefano
  • FASText: Efficient Unconstrained Scene Text Detector [pdf] - Michal Busta, Lukas Neumann, Jiri Matas
  • Context-Guided Diffusion for Label Propagation on Graphs [pdf] - Kwang In Kim, James Tompkin, Hanspeter Pfister, Christian Theobalt
  • Wide Baseline Stereo Matching With Convex Bounded Distortion Constraints [pdf] - Meirav Galun, Tal Amir, Tal Hassner, Ronen Basri, Yaron Lipman
Structural Kernel Learning for Large Scale Multiclass Object Co-Detection [pdf]
Zeeshan Hayder, Xuming He, Mathieu Salzmann

Abstract: 633 2634 2635 2636 2637 2638 2639 2640
Similar papers:
  • Love Thy Neighbors: Image Annotation by Exploiting Image Metadata [pdf] - Justin Johnson, Lamberto Ballan, Li Fei-Fei
  • On Linear Structure From Motion for Light Field Cameras [pdf] - Ole Johannsen, Antonin Sulc, Bastian Goldluecke
  • SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf] - Kensei Jo, Mohit Gupta, Shree K. Nayar
  • Pan-Sharpening With a Hyper-Laplacian Penalty [pdf] - Yiyong Jiang, Xinghao Ding, Delu Zeng, Yue Huang, John Paisley
  • Guiding the Long-Short Term Memory Model for Image Caption Generation [pdf] - Xu Jia, Efstratios Gavves, Basura Fernando, Tinne Tuytelaars
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification [pdf]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun

Abstract: Rectied activation units (rectiers) are essential for state-of-the-art neural networks. In this work, we study rectier neural networks for image classication from two aspects. First, we propose a Parametric Rectied Linear Unit (PReLU) that generalizes the traditional rectied unit. PReLU improves model tting with nearly zero extra computational cost and little overtting risk. Second, we derive a robust initialization method that particularly considers the rectier nonlinearities. This method enables us to train extremely deep rectied models directly from scratch and to investigate deeper or wider network architectures. Based on the learnable activation and advanced initialization, we achieve 4.94% top-5 test error on the ImageNet 2012 classication dataset. This is a 26% relative improvement over the ILSVRC 2014 winner (GoogLeNet, 6.66% [33]). To our knowledge, our result is the rst1 to surpass the reported human-level performance (5.1%, [26]) on this dataset.
Similar papers:
  • An Exploration of Parameter Redundancy in Deep Networks With Circulant Projections [pdf] - Yu Cheng, Felix X. Yu, Rogerio S. Feris, Sanjiv Kumar, Alok Choudhary, Shi-Fu Chang
  • Matrix Backpropagation for Deep Networks With Structured Layers [pdf] - Catalin Ionescu, Orestis Vantzos, Cristian Sminchisescu
  • Semantic Image Segmentation via Deep Parsing Network [pdf] - Ziwei Liu, Xiaoxiao Li, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Fast R-CNN [pdf] - Ross Girshick
  • Multi-Scale Recognition With DAG-CNNs [pdf] - Songfan Yang, Deva Ramanan
Oriented Object Proposals [pdf]
Shengfeng He, Rynson W.H. Lau

Abstract: In this paper, we propose a new approach to generate oriented object proposals (OOPs) to reduce the detection error caused by various orientations of the object. To this end, we propose to efficiently locate object regions according to pixelwise object probability, rather than measuring the objectness from a set of sampled windows. We formulate the proposal generation problem as a generative probabilistic model such that object proposals of different shapes (i.e., sizes and orientations) can be produced by locating the local maximum likelihoods. The new approach has three main advantages. First, it helps the object detector handle objects of different orientations. Second, as the shapes of the proposals may vary to fit the objects, the resulting proposals are tighter than the sampling windows with fixed sizes. Third, it avoids massive window sampling, and thereby reducing the number of proposals while maintaining a high recall. Experiments on the PASCAL VOC 2007 dataset show that the proposed OOP outperforms the stateof-the-art fast methods. Further experiments show that the rotation invariant property helps a class-specific object detector achieve better performance than the state-of-the-art proposal generation methods in either object rotation scenarios or general scenarios. Generating OOPs is very fast and takes only 0.5s per image.
Similar papers:
  • Box Aggregation for Proposal Decimation: Last Mile of Object Detection [pdf] - Shu Liu, Cewu Lu, Jiaya Jia
  • AttentionNet: Aggregating Weak Directions for Accurate Object Detection [pdf] - Donggeun Yoo, Sunggyun Park, Joon-Young Lee, Anthony S. Paek, In So Kweon
  • DeepProposal: Hunting Objects by Cascading Deep Convolutional Layers [pdf] - Amir Ghodrati, Ali Diba, Marco Pedersoli, Tinne Tuytelaars, Luc Van Gool
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • Cascaded Sparse Spatial Bins for Efficient and Effective Generic Object Detection [pdf] - David Novotny, Jiri Matas
Variational PatchMatch MultiView Reconstruction and Refinement [pdf]
Philipp Heise, Brian Jensen, Sebastian Klose, Alois Knoll

Abstract: In this work we propose a novel approach to the problem of multi-view stereo reconstruction. Building upon the previously proposed PatchMatch stereo and PM-Huber algorithm we introduce an extension to the multi-view scenario that employs an iterative refinement scheme. Our proposed approach uses an extended and robustified volumetric truncated signed distance function representation, which is advantageous for the fusion of refined depth maps and also for raycasting the current reconstruction estimation together with estimated depth normals into arbitrary camera views. We formulate the combined multi-view stereo reconstruction and refinement as a variational optimization problem. The newly introduced plane based smoothing term in the energy formulation is guided by the current reconstruction confidence and the image contents. Further we propose an extension of the PatchMatch scheme with an additional KLT step to avoid unnecessary sampling iterations. Improper camera poses are corrected by a direct image alignment step that performs robust outlier compensation by means of a recently proposed kernel lifting framework. To speed up the optimization of the variational formulation an adapted scheme is used for faster convergence.
Similar papers:
  • Global, Dense Multiscale Reconstruction for a Billion Points [pdf] - Benjamin Ummenhofer, Thomas Brox
  • Exploiting High Level Scene Cues in Stereo Reconstruction [pdf] - Simon Hadfield, Richard Bowden
  • Reflection Modeling for Passive Stereo [pdf] - Rahul Nair, Andrew Fitzgibbon, Daniel Kondermann, Carsten Rother
  • High Quality Structure From Small Motion for Rolling Shutter Cameras [pdf] - Sunghoon Im, Hyowon Ha, Gyeongmin Choe, Hae-Gon Jeon, Kyungdon Joo, In So Kweon
  • Massively Parallel Multiview Stereopsis by Surface Normal Diffusion [pdf] - Silvano Galliani, Katrin Lasinger, Konrad Schindler
Self-Calibration of Optical Lenses [pdf]
Michael Hirsch, Bernhard Scholkopf

Abstract: Even high-quality lenses suffer from optical aberrations, especially when used at full aperture. Furthermore, there are significant lens-to-lens deviations due to manufacturing tolerances, often rendering current software solutions like DxO, Lightroom, and PTLens insufficient as they dont adapt and only include generic lens blur models. We propose a method that enables the self-calibration of lenses from a natural image, or a set of such images. To this end we develop a machine learning framework that is able to exploit several recorded images and distills the available information into an accurate model of the considered lens.
Similar papers:
  • Blur-Aware Disparity Estimation From Defocus Stereo Images [pdf] - Ching-Hui Chen, Hui Zhou, Timo Ahonen
  • Video Super-Resolution via Deep Draft-Ensemble Learning [pdf] - Renjie Liao, Xin Tao, Ruiyu Li, Ziyang Ma, Jiaya Jia
  • Accurate Camera Calibration Robust to Defocus Using a Smartphone [pdf] - Hyowon Ha, Yunsu Bok, Kyungdon Joo, Jiyoung Jung, In So Kweon
  • Class-Specific Image Deblurring [pdf] - Saeed Anwar, Cong Phuoc Huynh, Fatih Porikli
  • Conditioned Regression Models for Non-Blind Single Image Super-Resolution [pdf] - Gernot Riegler, Samuel Schulter, Matthias Ruther, Horst Bischof
The HCI Stereo Metrics: Geometry-Aware Performance Analysis of Stereo Algorithms [pdf]
Katrin Honauer, Lena Maier-Hein, Daniel Kondermann

Abstract: Performance characterization of stereo methods is mandatory to decide which algorithm is useful for which application. Prevalent benchmarks mainly use the root mean squared error (RMS) with respect to ground truth disparity maps to quantify algorithm performance. We show that the RMS is of limited expressiveness for algorithm selection and introduce the HCI Stereo Metrics. These metrics assess stereo results by harnessing three semantic cues: depth discontinuities, planar surfaces, and fine geometric structures. For each cue, we extract the relevant set of pixels from existing ground truth. We then apply our evaluation functions to quantify characteristics such as edge fattening and surface smoothness. We demonstrate that our approach supports practitioners in selecting the most suitable algorithm for their application. Using the new Middlebury dataset, we show that rankings based on our metrics reveal specific algorithm strengths and weaknesses which are not quantified by existing metrics. We finally show how stacked bar charts and radar charts visually support multidimensional performance evaluation. An interactive stereo benchmark based on the proposed metrics and visualizations is available at: http://hci.iwr.uni-heidelberg.de/stereometrics
Similar papers:
  • MeshStereo: A Global Stereo Model With Mesh Alignment Regularization for View Interpolation [pdf] - Chi Zhang, Zhiwei Li, Yanhua Cheng, Rui Cai, Hongyang Chao, Yong Rui
  • Blur-Aware Disparity Estimation From Defocus Stereo Images [pdf] - Ching-Hui Chen, Hui Zhou, Timo Ahonen
  • StereoSnakes: Contour Based Consistent Object Extraction For Stereo Images [pdf] - Ran Ju, Tongwei Ren, Gangshan Wu
  • MAP Disparity Estimation Using Hidden Markov Trees [pdf] - Eric T. Psota, Jedrzej Kowalczuk, Mateusz Mittek, Lance C. Perez
  • A Deep Visual Correspondence Embedding Model for Stereo Matching Costs [pdf] - Zhuoyuan Chen, Xun Sun, Liang Wang, Yinan Yu, Chang Huang
Secrets of Matrix Factorization: Approximations, Numerics, Manifold Optimization and Random Restarts [pdf]
Je Hyeong Hong, Andrew Fitzgibbon

Abstract: Matrix factorization (or low-rank matrix completion) with missing data is a key computation in many computer vision and machine learning tasks, and is also related to a broader class of nonlinear optimization problems such as bundle adjustment. The problem has received much attention recently, with renewed interest in variable-projection approaches, yielding dramatic improvements in reliability and speed. However, on a wide class of problems, no one approach dominates, and because the various approaches have been derived in a multitude of different ways, it has been difficult to unify them. This paper provides a unified derivation of a number of recent approaches, so that similarities and differences are easily observed. We also present a simple meta-algorithm which wraps any existing algorithm, yielding 100% success rate on many standard datasets. Given 100% success, the focus of evaluation must turn to speed, as 100% success is trivially achieved if we do not care about speed. Again our unification allows a number of generic improvements applicable to all members of the family to be isolated, yielding a unified algorithm that outperforms our re-implementation of existing algorithms, which in some cases already outperform the original authors publicly available codes.
Similar papers:
  • A Versatile Learning-Based 3D Temporal Tracker: Scalable, Robust, Online [pdf] - David Joseph Tan, Federico Tombari, Slobodan Ilic, Nassir Navab
  • Semantic Component Analysis [pdf] - Calvin Murdock, Fernando De la Torre
  • Shape Interaction Matrix Revisited and Robustified: Efficient Subspace Clustering With Corrupted and Incomplete Data [pdf] - Pan Ji, Mathieu Salzmann, Hongdong Li
  • Convex Optimization With Abstract Linear Operators [pdf] - Steven Diamond, Stephen Boyd
  • Low-Rank Matrix Factorization Under General Mixture Noise Distributions [pdf] - Xiangyong Cao, Yang Chen, Qian Zhao, Deyu Meng, Yao Wang, Dong Wang, Zongben Xu
Regressive Tree Structured Model for Facial Landmark Localization [pdf]
Gee-Sern Hsu, Kai-Hsiang Chang, Shih-Chieh Huang

Abstract: Although the Tree Structured Model (TSM) is proven effective for solving face detection, pose estimation and landmark localization in an unified model, its sluggish runtime makes it unfavorable in practical applications, especially when dealing with cases of multiple faces. We propose the Regressive Tree Structure Model (RTSM) to improve the run-time speed and localization accuracy. The RTSM is composed of two component TSMs, the coarse TSM (cTSM) and the refined TSM (r-TSM), and a Bilateral Support Vector Regressor (BSVR). The c-TSM is built on the low-resolution octaves of samples so that it provides coarse but fast face detection. The r-TSM is built on the midresolution octaves so that it can locate the landmarks on the face candidates given by the c-TSM and improve precision. The r-TSM based landmarks are used in the forward BSVR as references to locate the dense set of landmarks, which are then used in the backward BSVR to relocate the landmarks with large localization errors. The forward and backward regression goes on iteratively until convergence. The performance of the RTSM is validated on three benchmark databases, the Multi-PIE, LFPW and AFW, and compared with the latest TSM to demonstrate its efficacy.
Similar papers:
  • Robust Facial Landmark Detection Under Significant Head Poses and Occlusion [pdf] - Yue Wu, Qiang Ji
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
  • Single Image Pop-Up From Discriminatively Learned Parts [pdf] - Menglong Zhu, Xiaowei Zhou, Kostas Daniilidis
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • Leveraging Datasets With Varying Annotations for Face Alignment via Deep Regression Network [pdf] - Jie Zhang, Meina Kan, Shiguang Shan, Xilin Chen
Unsupervised Domain Adaptation With Imbalanced Cross-Domain Data [pdf]
Tzu Ming Harry Hsu, Wei Yu Chen, Cheng-An Hou, Yao-Hung Hubert Tsai, Yi-Ren Yeh, Yu-Chiang Frank Wang

Abstract: We address a challenging unsupervised domain adaptation problem with imbalanced cross-domain data. For standard unsupervised domain adaptation, one typically obtains labeled data in the source domain and only observes unlabeled data in the target domain. However, most existing works do not consider the scenarios in which either the label numbers across domains are different, or the data in the source and/or target domains might be collected from multiple datasets. To address the aforementioned settings of imbalanced cross-domain data, we propose Closest Common Space Learning (CCSL) for associating such data with the capability of preserving label and structural information within and across domains. Experiments on multiple crossdomain visual classification tasks confirm that our method performs favorably against state-of-the-art approaches, especially when imbalanced cross-domain data are presented.
Similar papers:
  • Zero-Shot Learning via Semantic Similarity Embedding [pdf] - Ziming Zhang, Venkatesh Saligrama
  • Unsupervised Domain Adaptation for Zero-Shot Learning [pdf] - Elyor Kodirov, Tao Xiang, Zhenyong Fu, Shaogang Gong
  • Multi-View Domain Generalization for Visual Recognition [pdf] - Li Niu, Wen Li, Dong Xu
  • Simultaneous Deep Transfer Across Domains and Tasks [pdf] - Eric Tzeng, Judy Hoffman, Trevor Darrell, Kate Saenko
  • Bi-Shifting Auto-Encoder for Unsupervised Domain Adaptation [pdf] - Meina Kan, Shiguang Shan, Xilin Chen
Online Object Tracking With Proposal Selection [pdf]
Yang Hua, Karteek Alahari, Cordelia Schmid

Abstract: Tracking-by-detection approaches are some of the most successful object trackers in recent years. Their success is largely determined by the detector model they learn initially and then update over time. However, under challenging conditions where an object can undergo transformations, e.g., severe rotation, these methods are found to be lacking. In this paper, we address this problem by formulating it as a proposal selection task and making two contributions. The first one is introducing novel proposals estimated from the geometric transformations undergone by the object, and building a rich candidate set for predicting the object location. The second one is devising a novel selection strategy using multiple cues, i.e., detection score and edgeness score computed from state-of-the-art object edges and motion boundaries. We extensively evaluate our approach on the visual object tracking 2014 challenge and online tracking benchmark datasets, and show the best performance.
Similar papers:
  • Multi-Kernel Correlation Filter for Visual Tracking [pdf] - Ming Tang, Jiayi Feng
  • Learning Spatially Regularized Correlation Filters for Visual Tracking [pdf] - Martin Danelljan, Gustav Hager, Fahad Shahbaz Khan, Michael Felsberg
  • SOWP: Spatially Ordered and Weighted Patch Descriptor for Visual Tracking [pdf] - Han-Ul Kim, Dae-Youn Lee, Jae-Young Sim, Chang-Su Kim
  • A Versatile Learning-Based 3D Temporal Tracker: Scalable, Robust, Online [pdf] - David Joseph Tan, Federico Tombari, Slobodan Ilic, Nassir Navab
  • Understanding and Diagnosing Visual Tracking Systems [pdf] - Naiyan Wang, Jianping Shi, Dit-Yan Yeung, Jiaya Jia
Automatic Thumbnail Generation Based on Visual Representativeness and Foreground Recognizability [pdf]
Jingwei Huang, Huarong Chen, Bin Wang, Stephen Lin

Abstract: We present an automatic thumbnail generation technique based on two essential considerations: how well they visually represent the original photograph, and how well the foreground can be recognized after the cropping and downsizing steps of thumbnailing. These factors, while important for the image indexing purpose of thumbnails, have largely been ignored in previous methods, which instead are designed to highlight salient content while disregarding the effects of downsizing. We propose a set of image features for modeling these two considerations of thumbnails, and learn how to balance their relative effects on thumbnail generation through training on image pairs composed of photographs and their corresponding thumbnails created by an expert photographer. Experiments show the effectiveness of this approach on a variety of images, as well as its advantages over related techniques.
Similar papers:
  • Minimum Barrier Salient Object Detection at 80 FPS [pdf] - Jianming Zhang, Stan Sclaroff, Zhe Lin, Xiaohui Shen, Brian Price, Radomir Mech
  • A Data-Driven Metric for Comprehensive Evaluation of Saliency Models [pdf] - Jia Li, Changqun Xia, Yafei Song, Shu Fang, Xiaowu Chen
  • SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf] - Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao
  • HARF: Hierarchy-Associated Rich Features for Salient Object Detection [pdf] - Wenbin Zou, Nikos Komodakis
  • Cluster-Based Point Set Saliency [pdf] - Flora Ponjou Tasse, Jiri Kosinka, Neil Dodgson
Conditional High-Order Boltzmann Machine: A Supervised Learning Model for Relation Learning [pdf]
Yan Huang, Wei Wang, Liang Wang

Abstract: Relation learning is a fundamental operation in many computer vision tasks. Recently, high-order Boltzmann machine and its variants have exhibited the great power of modelling various data relation. However, most of them are unsupervised learning models which are not very discriminative and thus cannot server as a standalone solution to relation learning tasks. In this paper, we explore supervised learning algorithms and propose a new model named Conditional High-order Boltzmann Machine (CHBM), which can be directly used as a bilinear classifier to assign similarity scores for pairwise images. Then, to better deal with complex data relation, we propose a gated version of CHBM which untangles factors of variation by exploiting a set of latent variables to gate classification. We perform four-order tensor factorization for parameter reduction, and present two efficient supervised learning algorithms from the perspectives of being generative and discriminative, respectively. The experimental results of image transformation visualization, binary-way classification and face verification demonstrate that, by performing supervised learning, our models can greatly improve the performance.
Similar papers:
  • Personalized Age Progression With Aging Dictionary [pdf] - Xiangbo Shu, Jinhui Tang, Hanjiang Lai, Luoqi Liu, Shuicheng Yan
  • Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf] - Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Learning Social Relation Traits From Face Images [pdf] - Zhanpeng Zhang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Simultaneous Local Binary Feature Learning and Encoding for Face Recognition [pdf] - Jiwen Lu, Venice Erin Liong, Jie Zhou
Cross-Domain Image Retrieval With a Dual Attribute-Aware Ranking Network [pdf]
Junshi Huang, Rogerio S. Feris, Qiang Chen, Shuicheng Yan

Abstract: We address the problem of cross-domain image retrieval, considering the following practical application: given a user photo depicting a clothing image, our goal is to retrieve the same or attribute-similar clothing items from online shopping stores. This is a challenging problem due to the large discrepancy between online shopping images, usually taken in ideal lighting/pose/background conditions, and user photos captured in uncontrolled conditions. To address this problem, we propose a Dual Attribute-aware Ranking Network (DARN) for retrieval feature learning. More specically, DARN consists of two sub-networks, one for each domain, whose retrieval feature representations are driven by semantic attribute learning. We show that this attribute-guided learning is a key factor for retrieval accuracy improvement. In addition, to further align with the nature of the retrieval problem, we impose a triplet visual similarity constraint for learning to rank across the two subnetworks. Another contribution of our work is a large-scale dataset which makes the network learning feasible. We exploit customer review websites to crawl a large set of online shopping images and corresponding offline user photos with ne-grained clothing attributes, i.e., around 450,000 online shopping images and about 90,000 exact offline counterpart images of those online ones. All these images are collected from real-world consumer websites reflecting the diversity of the data modality, which makes this data
Similar papers:
  • Attribute-Graph: A Graph Based Approach to Image Ranking [pdf] - Nikita Prabhu, R. Venkatesh Babu
  • Learning Visual Clothing Style With Heterogeneous Dyadic Co-Occurrences [pdf] - Andreas Veit, Balazs Kovacs, Sean Bell, Julian McAuley, Kavita Bala, Serge Belongie
  • Localize Me Anywhere, Anytime: A Multi-Task Point-Retrieval Approach [pdf] - Guoyu Lu, Yan Yan, Li Ren, Jingkuan Song, Nicu Sebe, Chandra Kambhamettu
  • Multi-Label Cross-Modal Retrieval [pdf] - Viresh Ranjan, Nikhil Rasiwasia, C. V. Jawahar
  • Where to Buy It: Matching Street Clothing Photos in Online Shops [pdf] - M. Hadi Kiapour, Xufeng Han, Svetlana Lazebnik, Alexander C. Berg, Tamara L. Berg
Geometry-Aware Deep Transform [pdf]
Jiaji Huang, Qiang Qiu, Robert Calderbank, Guillermo Sapiro

Abstract: Many recent efforts have been devoted to designing sophisticated deep learning structures, obtaining revolutionary results on benchmark datasets. The success of these deep learning methods mostly relies on an enormous volume of labeled training samples to learn a huge number of parameters in a network; therefore, understanding the generalization ability of a learned deep network cannot be overlooked, especially when restricted to a small training set, which is the case for many applications. In this paper, we propose a novel deep learning objective formulation that unifies both the classification and metric learning criteria. We then introduce a geometry-aware deep transform to enable a non-linear discriminative and robust feature transform, which shows competitive performance on small training sets for both synthetic and real-world data. We further support the proposed framework with a formal (K, )-robustness analysis.
Similar papers:
  • Semantic Video Entity Linking Based on Visual Content and Metadata [pdf] - Yuncheng Li, Xitong Yang, Jiebo Luo
  • Conditional High-Order Boltzmann Machine: A Supervised Learning Model for Relation Learning [pdf] - Yan Huang, Wei Wang, Liang Wang
  • A Supervised Low-Rank Method for Learning Invariant Subspaces [pdf] - Farzad Siyahjani, Ranya Almohsen, Sinan Sabri, Gianfranco Doretto
  • Scalable Nonlinear Embeddings for Semantic Category-Based Image Retrieval [pdf] - Gaurav Sharma, Bernt Schiele
  • Simultaneous Local Binary Feature Learning and Encoding for Face Recognition [pdf] - Jiwen Lu, Venice Erin Liong, Jie Zhou
SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf]
Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao

Abstract: Saliency in Context (SALICON) is an ongoing effort that aims at understanding and predicting visual attention. Conventional saliency models typically rely on low-level image statistics to predict human fixations. While these models perform significantly better than chance, there is still a large gap between model prediction and human behavior. This gap is largely due to the limited capability of models in predicting eye fixations with strong semantic content, the so-called semantic gap. This paper presents a focused study to narrow the semantic gap with an architecture based on Deep Neural Network (DNN). It leverages the representational power of high-level semantics encoded in DNNs pretrained for object recognition. Two key components are fine-tuning the DNNs with an objective function based on the saliency evaluation metrics, and integrating information at different image scales. We compare our method with 14 saliency models on 6 public eye tracking benchmark datasets. Results demonstrate that our DNNs can automatically learn features for saliency prediction that surpass by a big margin the state-of-the-art. In addition, our model ranks top to date under all seven metrics on the MIT300 challenge set.
Similar papers:
  • Generic Promotion of Diffusion-Based Salient Object Detection [pdf] - Peng Jiang, Nuno Vasconcelos, Jingliang Peng
  • Learning to Predict Saliency on Face Images [pdf] - Mai Xu, Yun Ren, Zulin Wang
  • HARF: Hierarchy-Associated Rich Features for Salient Object Detection [pdf] - Wenbin Zou, Nikos Komodakis
  • Cluster-Based Point Set Saliency [pdf] - Flora Ponjou Tasse, Jiri Kosinka, Neil Dodgson
  • A Data-Driven Metric for Comprehensive Evaluation of Saliency Models [pdf] - Jia Li, Changqun Xia, Yafei Song, Shu Fang, Xiaowu Chen
The Middle Child Problem: Revisiting Parametric Min-Cut and Seeds for Object Proposals [pdf]
Ahmad Humayun, Fuxin Li, James M. Rehg

Abstract: Object proposals have recently fueled the progress in detection performance. These proposals aim to provide category-agnostic localizations for all objects in an image. One way to generate proposals is to perform parametric min-cuts over seed locations. This paper demonstrates that standard parametric-cut models are ineffective in obtaining medium-sized objects, which we refer to as the middle child problem. We propose a new energy minimization framework incorporating geodesic distances between segments which solves this problem. In addition, we introduce a new superpixel merging algorithm which can generate a small set of seeds that reliably cover a large number of objects of all sizes. We call our method POISE Proposals for Objects from Improved Seeds and Energies. POISE enables parametric min-cuts to reach their full potential. On PAS CAL VOC it generates 2,640 segments with an average overlap of 0.81, whereas the closest competing methods require more than 4,200 proposals to reach the same accuracy [24, 30]. We show detailed quantitative comparisons against 5 state-of-the-art methods on PASCAL VOC and Microsoft COCO segmentation challenges.
Similar papers:
  • Cascaded Sparse Spatial Bins for Efficient and Effective Generic Object Detection [pdf] - David Novotny, Jiri Matas
  • DeepProposal: Hunting Objects by Cascading Deep Convolutional Layers [pdf] - Amir Ghodrati, Ali Diba, Marco Pedersoli, Tinne Tuytelaars, Luc Van Gool
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • Boosting Object Proposals: From Pascal to COCO [pdf] - Jordi Pont-Tuset, Luc Van Gool
  • Learning to Combine Mid-Level Cues for Object Proposal Generation [pdf] - Tom Lee, Sanja Fidler, Sven Dickinson
A Projection Free Method for Generalized Eigenvalue Problem With a Nonsmooth Regularizer [pdf]
Seong Jae Hwang, Maxwell D. Collins, Sathya N. Ravi, Vamsi K. Ithapu, Nagesh Adluru, Sterling C. Johnson, Vikas Singh

Abstract: Eigenvalue problems are ubiquitous in computer vision, covering a very broad spectrum of applications ranging from estimation problems in multi-view geometry to image segmentation. Few other linear algebra problems have a more mature set of numerical routines available and many computer vision libraries leverage such tools extensively. However, the ability to call the underlying solver only as a black box can often become restrictive. Many human in the loop settings in vision frequently exploit supervision from an expert, to the extent that the user can be considered a subroutine in the overall system. In other cases, there is additional domain knowledge, side or even partial information that one may want to incorporate within the formulation. In general, regularizing a (generalized) eigenvalue problem with such side information remains difficult. Motivated by these needs, this paper presents an optimization scheme to solve generalized eigenvalue problems (GEP) involving a (nonsmooth) regularizer. We start from an alternative formulation of GEP where the feasibility set of the model involves the Stiefel manifold. The core of this paper presents an end to end stochastic optimization scheme for the resultant problem. We show how this general algorithm enables improved statistical analysis of brain imaging data where the regularizer is derived from other views of the disease pathology, involving clinical measurements and other image-derived representations.
Similar papers:
  • A Novel Sparsity Measure for Tensor Recovery [pdf] - Qian Zhao, Deyu Meng, Xu Kong, Qi Xie, Wenfei Cao, Yao Wang, Zongben Xu
  • Low-Rank Tensor Constrained Multiview Subspace Clustering [pdf] - Changqing Zhang, Huazhu Fu, Si Liu, Guangcan Liu, Xiaochun Cao
  • Beyond Gauss: Image-Set Matching on the Riemannian Manifold of PDFs [pdf] - Mehrtash Harandi, Mathieu Salzmann, Mahsa Baktashmotlagh
  • Interpolation on the Manifold of K Component GMMs [pdf] - Hyunwoo J. Kim, Nagesh Adluru, Monami Banerjee, Baba C. Vemuri, Vikas Singh
  • Semi-Supervised Normalized Cuts for Image Segmentation [pdf] - Selene E. Chew, Nathan D. Cahill
Structured Indoor Modeling [pdf]
Satoshi Ikehata, Hang Yang, Yasutaka Furukawa

Abstract: ing approaches exist. However, their output is either a pure polygon soup [30] or a set of planar patches [31]. This paper presents a novel 3D modeling framework that reconstructs an indoor scene as a structured model from panorama RGBD images. A scene geometry is represented as a graph, where nodes correspond to structural elements such as rooms, walls, and objects. The approach devises a structure grammar that defines how a scene graph can be manipulated. The grammar then drives a principled new reconstruction algorithm, where the grammar rules are sequentially applied to recover a structured model. The paper also proposes a new room segmentation algorithm and an offset-map reconstruction algorithm that are used in the framework and can enforce architectural shape priors far beyond existing state-of-the-art. The structured scene representation enables a variety of novel applications, ranging from indoor scene visualization, automated floorplan generation, Inverse-CAD, and more. We have tested our framework and algorithms on six synthetic and five real datasets with qualitative and quantitative evaluations. The source code and the data are available at the project website [15]. We establish a computational framework and algorithms for reconstructing structured indoor model from panorama RGBD images. We introduce a novel 3D model representation structure graph, whose nodes represent structural elements such as rooms, doors, and objects, and the edges represent their geom
Similar papers:
  • Semantic Segmentation of RGBD Images With Mutex Constraints [pdf] - Zhuo Deng, Sinisa Todorovic, Longin Jan Latecki
  • Procedural Editing of 3D Building Point Clouds [pdf] - Ilke Demir, Daniel G. Aliaga, Bedrich Benes
  • You Are Here: Mimicking the Human Thinking Process in Reading Floor-Plans [pdf] - Hang Chu, Dong Ki Kim, Tsuhan Chen
  • Learning Informative Edge Maps for Indoor Scene Layout Prediction [pdf] - Arun Mallya, Svetlana Lazebnik
  • Lost Shopping! Monocular Localization in Large Indoor Spaces [pdf] - Shenlong Wang, Sanja Fidler, Raquel Urtasun
High Quality Structure From Small Motion for Rolling Shutter Cameras [pdf]
Sunghoon Im, Hyowon Ha, Gyeongmin Choe, Hae-Gon Jeon, Kyungdon Joo, In So Kweon

Abstract: We present a practical 3D reconstruction method to obtain a high-quality dense depth map from narrow-baseline image sequences captured by commercial digital cameras, such as DSLRs or mobile phones. Depth estimation from small motion has gained interest as a means of various photographic editing, but important limitations present themselves in the form of depth uncertainty due to a narrow baseline and rolling shutter. To address these problems, we introduce a novel 3D reconstruction method from narrowbaseline image sequences that effectively handles the effects of a rolling shutter that occur from most of commercial digital cameras. Additionally, we present a depth propagation method to fill in the holes associated with the unknown pixels based on our novel geometric guidance model. Both qualitative and quantitative experimental results show that our new algorithm consistently generates better 3D depth maps than those by the state-of-the-art method.
Similar papers:
  • Model-Based Tracking at 300Hz Using Raw Time-of-Flight Observations [pdf] - Jan Stuhmer, Sebastian Nowozin, Andrew Fitzgibbon, Richard Szeliski, Travis Perry, Sunil Acharya, Daniel Cremers, Jamie Shotton
  • Variational PatchMatch MultiView Reconstruction and Refinement [pdf] - Philipp Heise, Brian Jensen, Sebastian Klose, Alois Knoll
  • Dense Continuous-Time Tracking and Mapping With Rolling Shutter RGB-D Cameras [pdf] - Christian Kerl, Jorg Stuckler, Daniel Cremers
  • Global Structure-From-Motion by Similarity Averaging [pdf] - Zhaopeng Cui, Ping Tan
  • Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf] - Naejin Kong, Michael J. Black
Matrix Backpropagation for Deep Networks With Structured Layers [pdf]
Catalin Ionescu, Orestis Vantzos, Cristian Sminchisescu

Abstract: Deep neural network architectures have recently produced excellent results in a variety of areas in artificial intelligence and visual recognition, well surpassing traditional shallow architectures trained using hand-designed features. The power of deep networks stems both from their ability to perform local computations followed by pointwise non-linearities over increasingly larger receptive fields, and from the simplicity and scalability of the gradient-descent training procedure based on backpropagation. An open problem is the inclusion of layers that perform global, structured matrix computations like segmentation (e.g. normalized cuts) or higher-order pooling (e.g. log-tangent space metrics defined over the manifold of symmetric positive definite matrices) while preserving the validity and efficiency of an end-to-end deep training framework. In this paper we propose a sound mathematical apparatus to formally integrate global structured computation into deep computation architectures. At the heart of our methodology is the development of the theory and practice of backpropagation that generalizes to the calculus of adjoint matrix variations. We perform segmentation experiments using the BSDS and MSCOCO benchmarks and demonstrate that deep networks relying on second-order pooling and normalized cuts layers, trained end-to-end using matrix backpropagation, outperform counterparts that do not take advantage of such global layers.
Similar papers:
  • Deep Fried Convnets [pdf] - Zichao Yang, Marcin Moczulski, Misha Denil, Nando de Freitas, Alex Smola, Le Song, Ziyu Wang
  • Fast R-CNN [pdf] - Ross Girshick
  • Learning The Structure of Deep Convolutional Networks [pdf] - Jiashi Feng, Trevor Darrell
  • An Exploration of Parameter Redundancy in Deep Networks With Circulant Projections [pdf] - Yu Cheng, Felix X. Yu, Rogerio S. Feris, Sanjiv Kumar, Alok Choudhary, Shi-Fu Chang
  • Multi-Scale Recognition With DAG-CNNs [pdf] - Songfan Yang, Deva Ramanan
Alternating Co-Quantization for Cross-Modal Hashing [pdf]
Go Irie, Hiroyuki Arai, Yukinobu Taniguchi

Abstract: This paper addresses the problem of unsupervised learning of binary hash codes for efficient cross-modal retrieval. Many unimodal hashing studies have proven that both similarity preservation of data and maintenance of quantization quality are essential for improving retrieval performance with binary hash codes. However, most existing crossmodal hashing methods mainly have focused on the former, and the latter still remains almost untouched. We propose a method to minimize the binary quantization errors, which is tailored to cross-modal hashing. Our approach, named Alternating Co-Quantization (ACQ), alternately seeks binary quantizers for each modality space with the help of connections to other modality data so that they give minimal quantization errors while preserving data similarities. ACQ can be coupled with various existing cross-modal dimension reduction methods such as Canonical Correlation Analysis (CCA) and substantially boosts their retrieval performance in the Hamming space. Extensive experiments demonstrate that ACQ can outperform several state-of-the-art methods, even when it is combined with simple CCA.
Similar papers:
  • Two Birds, One Stone: Jointly Learning Binary Code for Large-Scale Face Image Retrieval and Attributes Prediction [pdf] - Yan Li, Ruiping Wang, Haomiao Liu, Huajie Jiang, Shiguang Shan, Xilin Chen
  • Adaptive Hashing for Fast Similarity Search [pdf] - Fatih Cakir, Stan Sclaroff
  • An NMF Perspective on Binary Hashing [pdf] - Lopamudra Mukherjee, Sathya N. Ravi, Vamsi K. Ithapu, Tyler Holmes, Vikas Singh
  • Learning Binary Codes for Maximum Inner Product Search [pdf] - Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen
  • kNN Hashing With Factorized Neighborhood Representation [pdf] - Kun Ding, Chunlei Huo, Bin Fan, Chunhong Pan
Segment-Phrase Table for Semantic Segmentation, Visual Entailment and Paraphrasing [pdf]
Hamid Izadinia, Fereshteh Sadeghi, Santosh K. Divvala, Hannaneh Hajishirzi, Yejin Choi, Ali Farhadi

Abstract: We introduce Segment-Phrase Table (SPT), a large collection of bijective associations between textual phrases and their corresponding segmentations. Leveraging recent progress in object recognition and natural language semantics, we show how we can successfully build a highquality segment-phrase table using minimal human supervision. More importantly, we demonstrate the unique value unleashed by this rich bimodal resource, for both vision as well as natural language understanding. First, we show that fine-grained textual labels facilitate contextual reasoning that helps in satisfying semantic constraints across image segments. This feature enables us to achieve state-of-the-art segmentation results on benchmark datasets. Next, we show that the association of high-quality segmentations to textual phrases aids in richer semantic understanding and reasoning of these textual phrases. Leveraging this feature, we motivate the problem of visual entailment and visual paraphrasing, and demonstrate its utility on a large dataset.
Similar papers:
  • Common Subspace for Model and Similarity: Phrase Learning for Caption Generation From Images [pdf] - Yoshitaka Ushiku, Masataka Yamaguchi, Yusuke Mukuta, Tatsuya Harada
  • Flickr30k Entities: Collecting Region-to-Phrase Correspondences for Richer Image-to-Sentence Models [pdf] - Bryan A. Plummer, Liwei Wang, Chris M. Cervantes, Juan C. Caicedo, Julia Hockenmaier, Svetlana Lazebnik
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • BoxSup: Exploiting Bounding Boxes to Supervise Convolutional Networks for Semantic Segmentation [pdf] - Jifeng Dai, Kaiming He, Jian Sun
  • Joint Object and Part Segmentation Using Deep Learned Potentials [pdf] - Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille
Car That Knows Before You Do: Anticipating Maneuvers via Learning Temporal Driving Models [pdf]
Ashesh Jain, Hema S. Koppula, Bharad Raghavan, Shane Soh, Ashutosh Saxena

Abstract: Advanced Driver Assistance Systems (ADAS) have made driving safer over the last decade. They prepare vehicles for unsafe road conditions and alert drivers if they perform a dangerous maneuver. However, many accidents are unavoidable because by the time drivers are alerted, it is already too late. Anticipating maneuvers beforehand can alert drivers before they perform the maneuver and also give ADAS more time to avoid or prepare for the danger. In this work we anticipate driving maneuvers a few seconds before they occur. For this purpose we equip a car with cameras and a computing device to capture the driving context from both inside and outside of the car. We propose an Autoregressive Input-Output HMM to model the contextual information alongwith the maneuvers. We evaluate our approach on a diverse data set with 1180 miles of natural freeway and city driving and show that we can anticipate maneuvers 3.5 seconds before they occur with over 80% F1-score in real-time.
Similar papers:
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Temporal Perception and Prediction in Ego-Centric Video [pdf] - Yipin Zhou, Tamara L. Berg
  • Context Aware Active Learning of Activity Recognition Models [pdf] - Mahmudul Hasan, Amit K. Roy-Chowdhury
  • Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf] - Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic
Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf]
Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek

Abstract: The goal of this paper is to recognize actions in video without the need for examples. Different from traditional zero-shot approaches we do not demand the design and specification of attribute classifiers and class-to-attribute mappings to allow for transfer from seen classes to unseen classes. Our key contribution is objects2action, a semantic word embedding that is spanned by a skip-gram model of thousands of object categories. Action labels are assigned to an object encoding of unseen video based on a convex combination of action and object affinities. Our semantic embedding has three main characteristics to accommodate for the specifics of actions. First, we propose a mechanism to exploit multiple-word descriptions of actions and objects. Second, we incorporate the automated selection of the most responsive objects per action. And finally, we demonstrate how to extend our zero-shot approach to the spatio-temporal localization of actions in video. Experiments on four action datasets demonstrate the potential of our approach.
Similar papers:
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
  • Action Recognition by Hierarchical Mid-Level Action Elements [pdf] - Tian Lan, Yuke Zhu, Amir Roshan Zamir, Silvio Savarese
  • Action Localization in Videos Through Context Walk [pdf] - Khurram Soomro, Haroon Idrees, Mubarak Shah
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
Learning Image Representations Tied to Ego-Motion [pdf]
Dinesh Jayaraman, Kristen Grauman

Abstract: Understanding how images of objects and scenes behave in response to specific ego-motions is a crucial aspect of proper visual development, yet existing visual learning methods are conspicuously disconnected from the physical source of their images. We propose to exploit proprioceptive motor signals to provide unsupervised regularization in convolutional neural networks to learn visual representations from egocentric video. Specifically, we enforce that our learned features exhibit equivariance i.e. they respond predictably to transformations associated with distinct ego-motions. With three datasets, we show that our unsupervised feature learning approach significantly outperforms previous approaches on visual recognition and next-best-view prediction tasks. In the most challenging test, we show that features learned from video captured on an autonomous driving platform improve large-scale scene recognition in static images from a disjoint domain.
Similar papers:
  • Integrating Dashcam Views Through Inter-Video Mapping [pdf] - Hsin-I Chen, Yi-Ling Chen, Wei-Tse Lee, Fan Wang, Bing-Yu Chen
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
  • Unsupervised Object Discovery and Tracking in Video Collections [pdf] - Suha Kwak, Minsu Cho, Ivan Laptev, Jean Ponce, Cordelia Schmid
  • Unsupervised Learning of Spatiotemporally Coherent Metrics [pdf] - Ross Goroshin, Joan Bruna, Jonathan Tompson, David Eigen, Yann LeCun
  • Unsupervised Learning of Visual Representations Using Videos [pdf] - Xiaolong Wang, Abhinav Gupta
Complementary Sets of Shutter Sequences for Motion Deblurring [pdf]
Hae-Gon Jeon, Joon-Young Lee, Yudeog Han, Seon Joo Kim, In So Kweon

Abstract: In this paper, we present a novel multi-image motion deblurring method utilizing the coded exposure technique. The key idea of our work is to capture video frames with a set of complementary fluttering patterns to preserve spatial frequency details. We introduce an algorithm for generating a complementary set of binary sequences based on the modern communication theory and implement the coded exposure video system with an off-the-shelf machine vision camera. The effectiveness of our method is demonstrated on various challenging examples with quantitative and qualitative comparisons to other computational image capturing methods used for image deblurring.
Similar papers:
  • Conditioned Regression Models for Non-Blind Single Image Super-Resolution [pdf] - Gernot Riegler, Samuel Schulter, Matthias Ruther, Horst Bischof
  • Accurate Camera Calibration Robust to Defocus Using a Smartphone [pdf] - Hyowon Ha, Yunsu Bok, Kyungdon Joo, Jiyoung Jung, In So Kweon
  • Model-Based Tracking at 300Hz Using Raw Time-of-Flight Observations [pdf] - Jan Stuhmer, Sebastian Nowozin, Andrew Fitzgibbon, Richard Szeliski, Travis Perry, Sunil Acharya, Daniel Cremers, Jamie Shotton
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
  • Class-Specific Image Deblurring [pdf] - Saeed Anwar, Cong Phuoc Huynh, Fatih Porikli
Shape Interaction Matrix Revisited and Robustified: Efficient Subspace Clustering With Corrupted and Incomplete Data [pdf]
Pan Ji, Mathieu Salzmann, Hongdong Li

Abstract: The Shape Interaction Matrix (SIM) is one of the earliest approaches to performing subspace clustering (i.e., separating points drawn from a union of subspaces). In this paper, we revisit the SIM and reveal its connections to several recent subspace clustering methods. Our analysis lets us derive a simple, yet effective algorithm to robustify the SIM and make it applicable to realistic scenarios where the data is corrupted by noise. We justify our method by intuitive examples and the matrix perturbation theory. We then show how this approach can be extended to handle missing data, thus yielding an efficient and general subspace clustering algorithm. We demonstrate the benefits of our approach over state-of-the-art subspace clustering methods on several challenging motion segmentation and face clustering problems, where the data includes corrupted and missing measurements.
Similar papers:
  • Low-Rank Matrix Factorization Under General Mixture Noise Distributions [pdf] - Xiangyong Cao, Yang Chen, Qian Zhao, Deyu Meng, Yao Wang, Dong Wang, Zongben Xu
  • Deformable 3D Fusion: From Partial Dynamic 3D Observations to Complete 4D Models [pdf] - Weipeng Xu, Mathieu Salzmann, Yongtian Wang, Yue Liu
  • Low-Rank Tensor Constrained Multiview Subspace Clustering [pdf] - Changqing Zhang, Huazhu Fu, Si Liu, Guangcan Liu, Xiaochun Cao
  • Temporal Subspace Clustering for Human Motion Segmentation [pdf] - Sheng Li, Kang Li, Yun Fu
  • Multi-View Subspace Clustering [pdf] - Hongchang Gao, Feiping Nie, Xuelong Li, Heng Huang
Synthesizing Illumination Mosaics From Internet Photo-Collections [pdf]
Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm

Abstract: We propose a framework for the automatic creation of time-lapse mosaics of a given scene. We achieve this by leveraging the illumination variations captured in Internet photo-collections. In order to depict and characterize the illumination spectrum of a scene, our method relies on building discrete representations of the image appearance space through connectivity graphs defined over a pairwise image distance function. The smooth appearance transitions are found as the shortest path in the similarity graph among images, and robust image alignment is achieved by leveraging scene semantics, multi-view geometry, and image warping techniques. The attained results present an insightful and compact visualization of the scene illuminations captured in crowd-sourced imagery.
Similar papers:
  • Illumination Robust Color Naming via Label Propagation [pdf] - Yuanliu liu, Zejian Yuan, Badong Chen, Jianru Xue, Nanning Zheng
  • Convolutional Color Constancy [pdf] - Jonathan T. Barron
  • 3D Time-Lapse Reconstruction From Internet Photos [pdf] - Ricardo Martin-Brualla, David Gallup, Steven M. Seitz
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
  • Beyond White: Ground Truth Colors for Color Constancy Correction [pdf] - Dongliang Cheng, Brian Price, Scott Cohen, Michael S. Brown
Guiding the Long-Short Term Memory Model for Image Caption Generation [pdf]
Xu Jia, Efstratios Gavves, Basura Fernando, Tinne Tuytelaars

Abstract: In this work we focus on the problem of image caption generation. We propose an extension of the long short term memory (LSTM) model, which we coin gLSTM for short. In particular, we add semantic information extracted from the image as extra input to each unit of the LSTM block, with the aim of guiding the model towards solutions that are more tightly coupled to the image content. Additionally, we explore different length normalization strategies for beam search to avoid bias towards short sentences. On various benchmark datasets such as Flickr8K, Flickr30K and MS COCO, we obtain results that are on par with or better than the current state-of-the-art. LSTM a man in a blue shirt is sitting on a bench Semantic Information Retrieval Result: a man riding a bucking bull at a rodeo, a man is being thrown off a bull during a rodeo, bullrider at a rodeo riding a bull, four men in a rodeo with a bull bucking, three people wrestle with a bull at a rodeo, men on horses in the rodeo try to rope in a bull, a man wearing blue pants is riding a bucking bull, the cowboy prepares to lasso the bull Semantic embedding: gLSTM Guide a man in a red shirt stands on a bull
Similar papers:
  • Flickr30k Entities: Collecting Region-to-Phrase Correspondences for Richer Image-to-Sentence Models [pdf] - Bryan A. Plummer, Liwei Wang, Chris M. Cervantes, Juan C. Caicedo, Julia Hockenmaier, Svetlana Lazebnik
  • Common Subspace for Model and Similarity: Phrase Learning for Caption Generation From Images [pdf] - Yoshitaka Ushiku, Masataka Yamaguchi, Yusuke Mukuta, Tatsuya Harada
  • Multimodal Convolutional Neural Networks for Matching Image and Sentence [pdf] - Lin Ma, Zhengdong Lu, Lifeng Shang, Hang Li
  • Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books [pdf] - Yukun Zhu, Ryan Kiros, Rich Zemel, Ruslan Salakhutdinov, Raquel Urtasun, Antonio Torralba, Sanja Fidler
  • Sequence to Sequence - Video to Text [pdf] - Subhashini Venugopalan, Marcus Rohrbach, Jeffrey Donahue, Raymond Mooney, Trevor Darrell, Kate Saenko
Generic Promotion of Diffusion-Based Salient Object Detection [pdf]
Peng Jiang, Nuno Vasconcelos, Jingliang Peng

Abstract: In this work, we propose a generic scheme to promote any diffusion-based salient object detection algorithm by original ways to re-synthesize the diffusion matrix and construct the seed vector. We first make a novel analysis of the working mechanism of the diffusion matrix, which reveals the close relationship between saliency diffusion and spectral clustering. Following this analysis, we propose to re-synthesize the diffusion matrix from the most discriminative eigenvectors after adaptive re-weighting. Further, we propose to generate the seed vector based on the readily available diffusion maps, avoiding extra computation for color-based seed search. As a particular instance, we use inverse normalized Laplacian matrix as the original diffusion matrix and promote the corresponding salient object detection algorithm, which leads to superior performance as experimentally demonstrated.
Similar papers:
  • Minimum Barrier Salient Object Detection at 80 FPS [pdf] - Jianming Zhang, Stan Sclaroff, Zhe Lin, Xiaohui Shen, Brian Price, Radomir Mech
  • HARF: Hierarchy-Associated Rich Features for Salient Object Detection [pdf] - Wenbin Zou, Nikos Komodakis
  • Cluster-Based Point Set Saliency [pdf] - Flora Ponjou Tasse, Jiri Kosinka, Neil Dodgson
  • SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf] - Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao
  • Context-Guided Diffusion for Label Propagation on Graphs [pdf] - Kwang In Kim, James Tompkin, Hanspeter Pfister, Christian Theobalt
Pan-Sharpening With a Hyper-Laplacian Penalty [pdf]
Yiyong Jiang, Xinghao Ding, Delu Zeng, Yue Huang, John Paisley

Abstract: Pan-sharpening is the task of fusing spectral information in low resolution multispectral images with spatial information in a corresponding high resolution panchromatic image. In such approaches, there is a trade-off between spectral and spatial quality, as well as computational efficiency. We present a method for pan-sharpening in which a sparsity-promoting objective function preserves both spatial and spectral content, and is efficient to optimize. Our objective incorporates the 1/2 -norm in a way that can leverage recent computationally efficient methods, and 1 for which the alternating direction method of multipliers can be used. Additionally, our objective penalizes image gradients to enforce high resolution fidelity, and exploits the Fourier domain for further computational efficiency. Visual quality metrics demonstrate that our proposed objective function can achieve higher spatial and spectral resolution than several previous well-known methods with competitive computational efficiency.
Similar papers:
  • Learning Nonlinear Spectral Filters for Color Image Reconstruction [pdf] - Michael Moeller, Julia Diebold, Guy Gilboa, Daniel Cremers
  • RGB-Guided Hyperspectral Image Upsampling [pdf] - Hyeokhyen Kwon, Yu-Wing Tai
  • Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf] - Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • POP Image Fusion - Derivative Domain Image Fusion Without Reintegration [pdf] - Graham D. Finlayson, Alex E. Hayes
  • Hyperspectral Super-Resolution by Coupled Spectral Unmixing [pdf] - Charis Lanaras, Emmanuel Baltsavias, Konrad Schindler
SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf]
Kensei Jo, Mohit Gupta, Shree K. Nayar

Abstract: Sensors that measure their motion with respect to the surrounding environment (ego-motion sensors) can be broadly classified into two categories. First is inertial sensors such as accelerometers. In order to estimate position and velocity, these sensors integrate the measured acceleration, which often results in accumulation of large errors over time. Second, camera-based approaches such as SLAM that can measure position directly, but their performance depends on the surrounding scenes properties. These approaches cannot function reliably if the scene has low frequency textures or small depth variations. We present a novel ego-motion sensor called SpeDo that addresses these fundamental limitations. SpeDo is based on using coherent light sources and cameras with large defocus. Coherent light, on interacting with a scene, creates a high frequency interferometric pattern in the captured images, called speckle. We develop a theoretical model for speckle ow (motion of speckle as a function of sensor motion), and show that it is quasi-invariant to surrounding scenes properties. As a result, SpeDo can measure ego-motion (not derivative of motion) simply by estimating optical ow at a few image locations. We have built a low-cost and compact hardware prototype of SpeDo and demonstrated high precision 6 DOF ego-motion estimation for complex trajectories in scenarios where the scene properties are challenging (e.g., repeating or no texture) as well as unknown.
Similar papers:
  • Depth Selective Camera: A Direct, On-Chip, Programmable Technique for Depth Selectivity in Photography [pdf] - Ryuichi Tadano, Adithya Kumar Pediredla, Ashok Veeraraghavan
  • On Linear Structure From Motion for Light Field Cameras [pdf] - Ole Johannsen, Antonin Sulc, Bastian Goldluecke
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Extended Depth of Field Catadioptric Imaging Using Focal Sweep [pdf] - Ryunosuke Yokoya, Shree K. Nayar
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
On Linear Structure From Motion for Light Field Cameras [pdf]
Ole Johannsen, Antonin Sulc, Bastian Goldluecke

Abstract: We present a novel approach to relative pose estimation which is tailored to 4D light field cameras. From the relationships between scene geometry and light field structure and an analysis of the light field projection in terms of Plucker ray coordinates, we deduce a set of linear constraints on ray space correspondences between a pair of light field cameras. These can be applied to infer relative pose of the light field cameras and thus obtain a point cloud reconstruction of the scene. While the proposed method has interesting relationships to pose estimation for generalized cameras based on ray-to-ray correspondence, our experiments demonstrate that our approach is both more accurate and computationally more efficient. It also compares favorably to direct linear pose estimation based on aligning the 3D point clouds obtained by reconstructing depth for each individual light field. To further validate the method, we employ the pose estimates to merge light fields captured with hand-held consumer light field cameras into refocusable panoramas.
Similar papers:
  • SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf] - Kensei Jo, Mohit Gupta, Shree K. Nayar
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Photometric Stereo in a Scattering Medium [pdf] - Zak Murez, Tali Treibitz, Ravi Ramamoorthi, David Kriegman
  • Photometric Stereo With Small Angular Variations [pdf] - Jian Wang, Yasuyuki Matsushita, Boxin Shi, Aswin C. Sankaranarayanan
  • Nighttime Haze Removal With Glow and Multiple Light Colors [pdf] - Yu Li, Robby T. Tan, Michael S. Brown
Love Thy Neighbors: Image Annotation by Exploiting Image Metadata [pdf]
Justin Johnson, Lamberto Ballan, Li Fei-Fei

Abstract: Some images that are difficult to recognize on their own may become more clear in the context of a neighborhood of related images with similar social-network metadata. We build on this intuition to improve multilabel image annotation. Our model uses image metadata nonparametrically to generate neighborhoods of related images using Jaccard similarities, then uses a deep neural network to blend visual information from the image and its neighbors. Prior work typically models image metadata parametrically; in contrast, our nonparametric treatment allows our model to perform well even when the vocabulary of metadata changes between training and testing. We perform comprehensive experiments on the NUS-WIDE dataset, where we show that our model outperforms state-of-the-art methods for multilabel image annotation even when our model is forced to generalize to new types of metadata.
Similar papers:
  • Learning Semi-Supervised Representation Towards a Unified Optimization Framework for Semi-Supervised Learning [pdf] - Chun-Guang Li, Zhouchen Lin, Honggang Zhang, Jun Guo
  • Human Parsing With Contextualized Convolutional Neural Network [pdf] - Xiaodan Liang, Chunyan Xu, Xiaohui Shen, Jianchao Yang, Si Liu, Jinhui Tang, Liang Lin, Shuicheng Yan
  • ML-MG: Multi-Label Learning With Missing Labels Using a Mixed Graph [pdf] - Baoyuan Wu, Siwei Lyu, Bernard Ghanem
  • Constrained Convolutional Neural Networks for Weakly Supervised Segmentation [pdf] - Deepak Pathak, Philipp Krahenbuhl, Trevor Darrell
  • Relaxing From Vocabulary: Robust Weakly-Supervised Deep Learning for Vocabulary-Free Image Tagging [pdf] - Jianlong Fu, Yue Wu, Tao Mei, Jinqiao Wang, Hanqing Lu, Yong Rui
Automated Facial Trait Judgment and Election Outcome Prediction: Social Dimensions of Face [pdf]
Jungseock Joo, Francis F. Steen, Song-Chun Zhu

Abstract: The human face is a primary medium of human communication and a prominent source of information used to infer various attributes. In this paper, we study a fully automated system that can infer the perceived traits of a person from his face social dimensions, such as intelligence, honesty, and competence and how those traits can be used to predict the outcomes of real-world social events that involve long-term commitments, such as political elections, job hires, and marriage engagements. To this end, we propose a hierarchical model for enduring traits inferred from faces, incorporating high-level perceptions and intermediate-level attributes. We show that our trained model can successfully classify the outcomes of two important political events, only using the photographs of politicians faces. Firstly, it classifies the winners of a series of recent U.S. elections with the accuracy of 67.9% (Governors) and 65.5% (Senators). We also reveal that the different political offices require different types of preferred traits. Secondly, our model can categorize the political party affiliations of politicians, i.e., Democrats vs. Republicans, with the accuracy of 62.6% (male) and 60.1% (female). To the best of our knowledge, our paper is the first to use automated visual trait analysis to predict the outcomes of real-world social events. This approach is more scalable and objective than the prior behavioral studies, and opens for a range of new applications.
Similar papers:
  • Panoptic Studio: A Massively Multiview System for Social Motion Capture [pdf] - Hanbyul Joo, Hao Liu, Lei Tan, Lin Gui, Bart Nabbe, Iain Matthews, Takeo Kanade, Shohei Nobuhara, Yaser Sheikh
  • Two Birds, One Stone: Jointly Learning Binary Code for Large-Scale Face Image Retrieval and Attributes Prediction [pdf] - Yan Li, Ruiping Wang, Haomiao Liu, Huajie Jiang, Shiguang Shan, Xilin Chen
  • FaceDirector: Continuous Control of Facial Performance in Video [pdf] - Charles Malleson, Jean-Charles Bazin, Oliver Wang, Derek Bradley, Thabo Beeler, Adrian Hilton, Alexander Sorkine-Hornung
  • Learning Image and User Features for Recommendation in Social Networks [pdf] - Xue Geng, Hanwang Zhang, Jingwen Bian, Tat-Seng Chua
  • Learning Social Relation Traits From Face Images [pdf] - Zhanpeng Zhang, Ping Luo, Chen-Change Loy, Xiaoou Tang
Panoptic Studio: A Massively Multiview System for Social Motion Capture [pdf]
Hanbyul Joo, Hao Liu, Lei Tan, Lin Gui, Bart Nabbe, Iain Matthews, Takeo Kanade, Shohei Nobuhara, Yaser Sheikh

Abstract: We present an approach to capture the 3D structure and motion of a group of people engaged in a social interaction. The core challenges in capturing social interactions are: (1) occlusion is functional and frequent; (2) subtle motion needs to be measured over a space large enough to host a social group; and (3) human appearance and configuration variation is immense. The Panoptic Studio is a system organized around the thesis that social interactions should be measured through the perceptual integration of a large variety of view points. We present a modularized system designed around this principle, consisting of integrated structural, hardware, and software innovations. The system takes, as input, 480 synchronized video streams of multiple people engaged in social activities, and produces, as output, the labeled time-varying 3D structure of anatomical landmarks on individuals in the space. The algorithmic contributions include a hierarchical approach for generating skeletal trajectory proposals, and an optimization framework for skeletal reconstruction with trajectory reassociation.
Similar papers:
  • Dense Continuous-Time Tracking and Mapping With Rolling Shutter RGB-D Cameras [pdf] - Christian Kerl, Jorg Stuckler, Daniel Cremers
  • Action Detection by Implicit Intentional Motion Clustering [pdf] - Wei Chen, Jason J. Corso
  • Uncovering Interactions and Interactors: Joint Estimation of Head, Body Orientation and F-Formations From Surveillance Videos [pdf] - Elisa Ricci, Jagannadan Varadarajan, Ramanathan Subramanian, Samuel Rota Bulo, Narendra Ahuja, Oswald Lanz
  • Motion Trajectory Segmentation via Minimum Cost Multicuts [pdf] - Margret Keuper, Bjoern Andres, Thomas Brox
  • Unsupervised Trajectory Clustering via Adaptive Multi-Kernel-Based Shrinkage [pdf] - Hongteng Xu, Yang Zhou, Weiyao Lin, Hongyuan Zha
Pose-Invariant 3D Face Alignment [pdf]
Amin Jourabloo, Xiaoming Liu

Abstract: Face alignment aims to estimate the locations of a set of landmarks for a given image. This problem has received much attention as evidenced by the recent advancement in both the methodology and performance. However, most of the existing works neither explicitly handle face images with arbitrary poses, nor perform large-scale experiments on non-frontal and profile face images. In order to address these limitations, this paper proposes a novel face alignment algorithm that estimates both 2D and 3D landmarks and their 2D visibilities for a face image with an arbitrary pose. By integrating a 3D point distribution model, a cascaded coupled-regressor approach is designed to estimate both the camera projection matrix and the 3D landmarks. Furthermore, the 3D model also allows us to automatically estimate the 2D landmark visibilities via surface normal. We use a substantially larger collection of all-pose face images to evaluate our algorithm and demonstrate superior performances than the state-of-the-art methods.
Similar papers:
  • Regressive Tree Structured Model for Facial Landmark Localization [pdf] - Gee-Sern Hsu, Kai-Hsiang Chang, Shih-Chieh Huang
  • Robust Statistical Face Frontalization [pdf] - Christos Sagonas, Yannis Panagakis, Stefanos Zafeiriou, Maja Pantic
  • Leveraging Datasets With Varying Annotations for Face Alignment via Deep Regression Network [pdf] - Jie Zhang, Meina Kan, Shiguang Shan, Xilin Chen
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
StereoSnakes: Contour Based Consistent Object Extraction For Stereo Images [pdf]
Ran Ju, Tongwei Ren, Gangshan Wu

Abstract: Consistent object extraction plays an essential role for stereo image editing with the population of stereoscopic 3D media. Most previous methods perform segmentation on entire images for both views using dense stereo correspondence constraints. We find that for such kind of methods the computation is highly redundant since the two views are near-duplicate. Besides, the consistency may be violated due to the imperfectness of current stereo matching algorithms. In this paper, we propose a contour based method which searches for consistent object contours instead of regions. It integrates both stereo correspondence and object boundary constraints into an energy minimization framework. The proposed method has several advantages compared to previous works. First, the searching space is restricted in object boundaries thus the efficiency significantly improved. Second, the discriminative power of object contours results in a more consistent segmentation. Furthermore, the proposed method can effortlessly extend existing single-image segmentation methods to work in stereo scenarios. The experiment on the Adobe benchmark shows superior extraction accuracy and significant improvement of efficiency of our method to state-of-the-art. We also demonstrate in a few applications how our method can be used as a basic tool for stereo image editing.
Similar papers:
  • Depth Map Estimation and Colorization of Anaglyph Images Using Local Color Prior and Reverse Intensity Distribution [pdf] - Williem, Ramesh Raskar, In Kyu Park
  • Interactive Visual Hull Refinement for Specular and Transparent Object Surface Reconstruction [pdf] - Xinxin Zuo, Chao Du, Sen Wang, Jiangbin Zheng, Ruigang Yang
  • Contour Flow: Middle-Level Motion Estimation by Combining Motion Segmentation and Contour Alignment [pdf] - Huijun Di, Qingxuan Shi, Feng Lv, Ming Qin, Yao Lu
  • The HCI Stereo Metrics: Geometry-Aware Performance Analysis of Stereo Algorithms [pdf] - Katrin Honauer, Lena Maier-Hein, Daniel Kondermann
  • A Deep Visual Correspondence Embedding Model for Stereo Matching Costs [pdf] - Zhuoyuan Chen, Xun Sun, Liang Wang, Yinan Yu, Chang Huang
Joint Fine-Tuning in Deep Neural Networks for Facial Expression Recognition [pdf]
Heechul Jung, Sihaeng Lee, Junho Yim, Sunjeong Park, Junmo Kim

Abstract: Temporal information has useful features for recognizing facial expressions. However, to manually design useful features requires a lot of effort. In this paper, to reduce this effort, a deep learning technique, which is regarded as a tool to automatically extract useful features from raw data, is adopted. Our deep network is based on two different models. The first deep network extracts temporal appearance features from image sequences, while the other deep network extracts temporal geometry features from temporal facial landmark points. These two models are combined using a new integration method in order to boost the performance of the facial expression recognition. Through several experiments, we show that the two models cooperate with each other. As a result, we achieve superior performance to other state-of-the-art methods in the CK+ and Oulu-CASIA databases. Furthermore, we show that our new integration method gives more accurate results than traditional methods, such as a weighted summation and a feature concatenation method.
Similar papers:
  • Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf] - Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic
  • Robust Facial Landmark Detection Under Significant Head Poses and Occlusion [pdf] - Yue Wu, Qiang Ji
  • Pairwise Conditional Random Forests for Facial Expression Recognition [pdf] - Arnaud Dapogny, Kevin Bailly, Severine Dubuisson
  • FaceDirector: Continuous Control of Facial Performance in Video [pdf] - Charles Malleson, Jean-Charles Bazin, Oliver Wang, Derek Bradley, Thabo Beeler, Adrian Hilton, Alexander Sorkine-Hornung
  • From Emotions to Action Units With Hidden and Semi-Hidden-Task Learning [pdf] - Adria Ruiz, Joost Van de Weijer, Xavier Binefa
Real-Time Pose Estimation Piggybacked on Object Detection [pdf]
Roman Juranek, Adam Herout, Marketa Dubska, Pavel Zemcik

Abstract: H = 25.1 We present an object detector coupled with pose estimation directly in a single compact and simple model, where the detector shares extracted image features with the pose estimator. The output of the classification of each candidate window consists of both object score and likelihood map of poses. This extension introduces negligible overhead during detection so that the detector is still capable of real time operation. We evaluated the proposed approach on the problem of vehicle detection. We used existing datasets with viewpoint/pose annotation (WCVP, 3D objects, KITTI). Besides that, we collected a new traffic surveillance dataset COD20k which fills certain gaps of the existing datasets and we make it public. The experimental results show that the proposed approach is comparable with state-of-the-art approaches in terms of accuracy, but it is considerably faster easily operating in real time (Matlab with C++ code). The source codes and the collected COD20k dataset are made public along with the paper.
Similar papers:
  • Unsupervised Generation of a Viewpoint Annotated Car Dataset From Videos [pdf] - Nima Sedaghat, Thomas Brox
  • Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf] - Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang
  • Continuous Pose Estimation With a Spatial Ensemble of Fisher Regressors [pdf] - Michele Fenzi, Laura Leal-Taixe, Jorn Ostermann, Tinne Tuytelaars
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • Pose Induction for Novel Object Categories [pdf] - Shubham Tulsiani, Joao Carreira, Jitendra Malik
Polarized 3D: High-Quality Depth Sensing With Polarization Cues [pdf]
Achuta Kadambi, Vage Taamazyan, Boxin Shi, Ramesh Raskar

Abstract: Coarse depth maps can be enhanced by using the shape information from polarization cues. We propose a framework to combine surface normals from polarization (hereafter polarization normals) with an aligned depth map. Polarization normals have not been used for depth enhancement before. This is because polarization normals suffer from physics-based artifacts, such as azimuthal ambiguity, refractive distortion and fronto-parallel signal degradation. We propose a framework to overcome these key challenges, allowing the benefits of polarization to be used to enhance depth maps. Our results demonstrate improvement with respect to state-of-the-art 3D reconstruction techniques. Passive capture: assuming light incident on an object is unpolarized, the surface normals can be obtained by rotating a polarizer at the imaging sensor. achoo@mit.edu SUTD Robustness to diffuse interreflections: unlike SfS and PS, diffuse interreflections do not significantly corrupt the estimated shape. Material invariant capture: the physics of the shape from polarization problem hold for materials ranging from dielectrics to metals to translucent objects. Lighting robust capture: if the incident light is unpolarized shape estimation is robust and can be conducted indoors, outdoors, or under patterned illumination. However, obtaining surface normals through polarization is not yet a mature technique. The obtained normals are drastically distorted. Specific open problems [24, 3] include: 1. Ambi
Similar papers:
  • Peeking Template Matching for Depth Extension [pdf] - Simon Korman, Eyal Ofek, Shai Avidan
  • Reflection Modeling for Passive Stereo [pdf] - Rahul Nair, Andrew Fitzgibbon, Daniel Kondermann, Carsten Rother
  • Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf] - Naejin Kong, Michael J. Black
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Predicting Depth, Surface Normals and Semantic Labels With a Common Multi-Scale Convolutional Architecture [pdf] - David Eigen, Rob Fergus
Bi-Shifting Auto-Encoder for Unsupervised Domain Adaptation [pdf]
Meina Kan, Shiguang Shan, Xilin Chen

Abstract: In many real-world applications, the domain of model learning (referred as source domain) is usually inconsistent with or even different from the domain of testing (referred as target domain), which makes the learnt model degenerate in target domain, i.e., the test domain. To alleviate the discrepancy between source and target domains, we propose a domain adaptation method, named as Bi-shifting Auto-Encoder network (BAE). The proposed BAE attempts to shift source domain samples to target domain, and also shift the target domain samples to source domain. The non-linear transformation of BAE ensures the feasibility of shifting between domains, and the distribution consistency between the shifted domain and the desirable domain is constrained by sparse reconstruction between them. As a result, the shifted source domain is supervised and follows similar distribution as target domain. Therefore, any supervised method can be applied on the shifted source domain to train a classifier for classification in target domain. The proposed method is evaluated on three domain adaptation scenarios of face recognition, i.e., domain adaptation across view angle, ethnicity, and imaging sensor, and the promising results demonstrate that our proposed BAE can shift samples between domains and thus effectively deal with the domain discrepancy.
Similar papers:
  • Zero-Shot Learning via Semantic Similarity Embedding [pdf] - Ziming Zhang, Venkatesh Saligrama
  • Unsupervised Domain Adaptation for Zero-Shot Learning [pdf] - Elyor Kodirov, Tao Xiang, Zhenyong Fu, Shaogang Gong
  • Multi-View Domain Generalization for Visual Recognition [pdf] - Li Niu, Wen Li, Dong Xu
  • Unsupervised Domain Adaptation With Imbalanced Cross-Domain Data [pdf] - Tzu Ming Harry Hsu, Wei Yu Chen, Cheng-An Hou, Yao-Hung Hubert Tsai, Yi-Ren Yeh, Yu-Chiang Frank Wang
  • Simultaneous Deep Transfer Across Domains and Tasks [pdf] - Eric Tzeng, Judy Hoffman, Trevor Darrell, Kate Saenko
Amodal Completion and Size Constancy in Natural Scenes [pdf]
Abhishek Kar, Shubham Tulsiani, Joao Carreira, Jitendra Malik

Abstract: We consider the problem of enriching current object detection systems with veridical object sizes and relative depth estimates from a single image. There are several technical challenges to this, such as occlusions, lack of calibration data and the scale ambiguity between object size and distance. These have not been addressed in full generality in previous work. Here we propose to tackle these issues by building upon advances in object recognition and using recently created large-scale datasets. We first introduce the task of amodal bounding box completion, which aims to infer the the full extent of the object instances in the image. We then propose a probabilistic framework for learning category-specific object size distributions from available annotations and leverage these in conjunction with amodal completions to infer veridical sizes of objects in novel images. Finally, we introduce a focal length prediction approach that exploits scene recognition to overcome inherent scale ambiguities and demonstrate qualitative results on challenging real-world scenes. Figure 1: Perceiving the veridical size of objects in realistic scenes, from a single image, requires disentangling size and depth, being able to compensate for occlusions and to determine intrinsic camera parameters. We tackle all three of these problems, leveraging recent developments in object recognition and large annotated object and scene datasets.
Similar papers:
  • Depth Recovery From Light Field Using Focal Stack Symmetry [pdf] - Haiting Lin, Can Chen, Sing Bing Kang, Jingyi Yu
  • BoxSup: Exploiting Bounding Boxes to Supervise Convolutional Networks for Semantic Segmentation [pdf] - Jifeng Dai, Kaiming He, Jian Sun
  • Oriented Object Proposals [pdf] - Shengfeng He, Rynson W.H. Lau
  • AttentionNet: Aggregating Weak Directions for Accurate Object Detection [pdf] - Donggeun Yoo, Sunggyun Park, Joon-Young Lee, Anthony S. Paek, In So Kweon
  • Extended Depth of Field Catadioptric Imaging Using Focal Sweep [pdf] - Ryunosuke Yokoya, Shree K. Nayar
Image Matting With KL-Divergence Based Sparse Sampling [pdf]
Levent Karacan, Aykut Erdem, Erkut Erdem

Abstract: Previous sampling-based image matting methods typically rely on certain heuristics in collecting representative samples from known regions, and thus their performance deteriorates if the underlying assumptions are not satisfied. To alleviate this, in this paper we take an entirely new approach and formulate sampling as a sparse subset selection problem where we propose to pick a small set of candidate samples that best explains the unknown pixels. Moreover, we describe a new distance measure for comparing two samples which is based on KL-divergence between the distributions of features extracted in the vicinity of the samples. Using a standard benchmark dataset for image matting, we demonstrate that our approach provides more accurate results compared with the state-of-the-art methods.
Similar papers:
  • Active Transfer Learning With Zero-Shot Priors: Reusing Past Datasets for Future Tasks [pdf] - Efstratios Gavves, Thomas Mensink, Tatiana Tommasi, Cees G. M. Snoek, Tinne Tuytelaars
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
  • Simultaneous Foreground Detection and Classification With Hybrid Features [pdf] - Jaemyun Kim, Adin Ramirez Rivera, Byungyong Ryu, Oksam Chae
  • Frequency-Based Environment Matting by Compressive Sensing [pdf] - Yiming Qian, Minglun Gong, Yee-Hong Yang
  • Video Matting via Sparse and Low-Rank Representation [pdf] - Dongqing Zou, Xiaowu Chen, Guangying Cao, Xiaogang Wang
Person Re-Identification With Discriminatively Trained Viewpoint Invariant Dictionaries [pdf]
Srikrishna Karanam, Yang Li, Richard J. Radke

Abstract: This paper introduces a new approach to address the person re-identification problem in cameras with nonoverlapping fields of view. Unlike previous approaches that learn Mahalanobis-like distance metrics in some transformed feature space, we propose to learn a dictionary that is capable of discriminatively and sparsely encoding features representing different people. Our approach directly addresses two key challenges in person re-identification: viewpoint variations and discriminability. First, to tackle viewpoint and associated appearance changes, we learn a single dictionary to represent both gallery and probe images in the training phase. We then discriminatively train the dictionary by enforcing explicit constraints on the associated sparse representations of the feature vectors. In the testing phase, we re-identify a probe image by simply determining the gallery image that has the closest sparse representation to that of the probe image in the Euclidean sense. Extensive performance evaluations on three publicly available multi-shot re-identification datasets demonstrate the advantages of our algorithm over several state-of-theart dictionary learning, temporal sequence matching, and spatial appearance and metric learning based techniques.
Similar papers:
  • Person Re-Identification Ranking Optimisation by Discriminant Context Information Analysis [pdf] - Jorge Garcia, Niki Martinel, Christian Micheloni, Alfredo Gardel
  • Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf] - Yuhui Quan, Yan Huang, Hui Ji
  • Partial Person Re-Identification [pdf] - Wei-Shi Zheng, Xiang Li, Tao Xiang, Shengcai Liao, Jianhuang Lai, Shaogang Gong
  • Multi-Task Learning With Low Rank Attribute Embedding for Person Re-Identification [pdf] - Chi Su, Fan Yang, Shiliang Zhang, Qi Tian, Larry S. Davis, Wen Gao
  • Conformal and Low-Rank Sparse Representation for Image Restoration [pdf] - Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng
On the Visibility of Point Clouds [pdf]
Sagi Katz, Ayellet Tal

Abstract: Is it possible to determine the visible subset of points directly from a given point cloud? Interestingly, in [7] it was shown that this is indeed the casedespite the fact that points cannot occlude each other, this task can be performed without surface reconstruction or normal estimation. The operator is very simpleit first transforms the points to a new domain and then constructs the convex hull in that domain. Points that lie on the convex hull of the transformed set of points are the images of the visible points. This operator found numerous applications in computer vision, including face reconstruction, keypoint detection, finding the best viewpoints, reduction of points, and many more. The current paper addresses a fundamental question: What properties should a transformation function satisfy, in order to be utilized in this operator? We show that three such properties are sufficientthe sign of the function, monotonicity, and a condition regarding the functions parameter. The correctness of an algorithm that satisfies these three properties is proved. Finally, we show an interesting application of the operatorassignment of visibility-confidence score. This feature is missing from previous approaches, where a binary yes/no visibility is determined. This score can be utilized in various applications; we illustrate its use in view-dependent curvature estimation.
Similar papers:
  • Convex Optimization With Abstract Linear Operators [pdf] - Steven Diamond, Stephen Boyd
  • Component-Wise Modeling of Articulated Objects [pdf] - Valsamis Ntouskos, Marta Sanzari, Bruno Cafaro, Federico Nardi, Fabrizio Natola, Fiora Pirri, Manuel Ruiz
  • Thin Structure Estimation With Curvature Regularization [pdf] - Dmitrii Marin, Yuchen Zhong, Maria Drangova, Yuri Boykov
  • Peeking Template Matching for Depth Extension [pdf] - Simon Korman, Eyal Ofek, Shai Avidan
  • A Versatile Scene Model With Differentiable Visibility Applied to Generative Pose Estimation [pdf] - Helge Rhodin, Nadia Robertini, Christian Richardt, Hans-Peter Seidel, Christian Theobalt
Active One-Shot Scan for Wide Depth Range Using a Light Field Projector Based on Coded Aperture [pdf]
Hiroshi Kawasaki, Satoshi Ono, Yuki Horita, Yuki Shiba, Ryo Furukawa, Shinsaku Hiura

Abstract: The central projection model commonly used to model cameras as well as projectors, results in similar advantages and disadvantages in both types of system. Considering the case of active stereo systems using a projector and camera setup, a central projection model creates several problems; among them, narrow depth range and necessity of wide baseline are crucial. In the paper, we solve the problems by introducing a light field projector, which can project a depth-dependent pattern. The light field projector is realized by attaching a coded aperture with a high frequency mask in front of the lens of the video projector, which also projects a high frequency pattern. Because the light field projector cannot be approximated by a thin lens model and a precise calibration method is not established yet, an image-based approach is proposed to apply a stereo technique to the system. Although image-based techniques usually require a large database and often imply heavy computational costs, we propose a hierarchical approach and a feature-based search for solution. In the experiments, it is confirmed that our method can accurately recover the dense shape of curved and textured objects for a wide range of depths from a single captured image.
Similar papers:
  • Extended Depth of Field Catadioptric Imaging Using Focal Sweep [pdf] - Ryunosuke Yokoya, Shree K. Nayar
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Photometric Stereo With Small Angular Variations [pdf] - Jian Wang, Yasuyuki Matsushita, Boxin Shi, Aswin C. Sankaranarayanan
  • SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf] - Kensei Jo, Mohit Gupta, Shree K. Nayar
  • On Linear Structure From Motion for Light Field Cameras [pdf] - Ole Johannsen, Antonin Sulc, Bastian Goldluecke
PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization [pdf]
Alex Kendall, Matthew Grimes, Roberto Cipolla

Abstract: We present a robust and real-time monocular six degree of freedom relocalization system. Our system trains a convolutional neural network to regress the 6-DOF camera pose from a single RGB image in an end-to-end manner with no need of additional engineering or graph optimisation. The algorithm can operate indoors and outdoors in real time, taking 5ms per frame to compute. It obtains approximately 2m and 3 accuracy for large scale outdoor scenes and 0.5m and 5 accuracy indoors. This is achieved using an efficient 23 layer deep convnet, demonstrating that convnets can be used to solve complicated out of image plane regression problems. This was made possible by leveraging transfer learning from large scale classification data. We show that the PoseNet localizes from high level features and is robust to difficult lighting, motion blur and different camera intrinsics where point based SIFT registration fails. Furthermore we show how the pose feature that is produced generalizes to other scenes allowing us to regress pose with only a few dozen training examples.
Similar papers:
  • Real-Time Pose Estimation Piggybacked on Object Detection [pdf] - Roman Juranek, Adam Herout, Marketa Dubska, Pavel Zemcik
  • Semantic Pose Using Deep Networks Trained on Synthetic RGB-D [pdf] - Jeremie Papon, Markus Schoeler
  • Flowing ConvNets for Human Pose Estimation in Videos [pdf] - Tomas Pfister, James Charles, Andrew Zisserman
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • Robust Optimization for Deep Regression [pdf] - Vasileios Belagiannis, Christian Rupprecht, Gustavo Carneiro, Nassir Navab
Dense Continuous-Time Tracking and Mapping With Rolling Shutter RGB-D Cameras [pdf]
Christian Kerl, Jorg Stuckler, Daniel Cremers

Abstract: We propose a dense continuous-time tracking and mapping method for RGB-D cameras. We parametrize the camera trajectory using continuous B-splines and optimize the trajectory through dense, direct image alignment. Our method also directly models rolling shutter in both RGB and depth images within the optimization, which improves tracking and reconstruction quality for low-cost CMOS sensors. Using a continuous trajectory representation has a number of advantages over a discrete-time representation (e.g. camera poses at the frame interval). With splines, less variables need to be optimized than with a discrete representation, since the trajectory can be represented with fewer control points than frames. Splines also naturally include smoothness constraints on derivatives of the trajectory estimate. Finally, the continuous trajectory representation allows to compensate for rolling shutter effects, since a pose estimate is available at any exposure time of an image. Our approach demonstrates superior quality in tracking and reconstruction compared to approaches with discrete-time or global shutter assumptions.
Similar papers:
  • A Novel Representation of Parts for Accurate 3D Object Detection and Tracking in Monocular Images [pdf] - Alberto Crivellaro, Mahdi Rad, Yannick Verdie, Kwang Moo Yi, Pascal Fua, Vincent Lepetit
  • Panoptic Studio: A Massively Multiview System for Social Motion Capture [pdf] - Hanbyul Joo, Hao Liu, Lei Tan, Lin Gui, Bart Nabbe, Iain Matthews, Takeo Kanade, Shohei Nobuhara, Yaser Sheikh
  • Model-Based Tracking at 300Hz Using Raw Time-of-Flight Observations [pdf] - Jan Stuhmer, Sebastian Nowozin, Andrew Fitzgibbon, Richard Szeliski, Travis Perry, Sunil Acharya, Daniel Cremers, Jamie Shotton
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
  • High Quality Structure From Small Motion for Rolling Shutter Cameras [pdf] - Sunghoon Im, Hyowon Ha, Gyeongmin Choe, Hae-Gon Jeon, Kyungdon Joo, In So Kweon
Efficient Decomposition of Image and Mesh Graphs by Lifted Multicuts [pdf]
Margret Keuper, Evgeny Levinkov, Nicolas Bonneel, Guillaume Lavoue, Thomas Brox, Bjorn Andres

Abstract: Formulations of the Image Decomposition Problem [8] as a Multicut Problem (MP) w.r.t. a superpixel graph have received considerable attention. In contrast, instances of the MP w.r.t. a pixel grid graph have received little attention, firstly, because the MP is NP-hard and instances w.r.t. a pixel grid graph are hard to solve in practice, and, secondly, due to the lack of long-range terms in the objective function of the MP. We propose a generalization of the MP with longrange terms (LMP). We design and implement two efficient algorithms (primal feasible heuristics) for the MP and LMP which allow us to study instances of both problems w.r.t. the pixel grid graphs of the images in the BSDS-500 benchmark [8]. The decompositions we obtain do not differ significantly from the state of the art, suggesting that the LMP is a competitive formulation of the Image Decomposition Problem. To demonstrate the generality of the LMP, we apply it also to the Mesh Decomposition Problem posed by the Princeton benchmark [16], obtaining state-of-the-art decompositions.
Similar papers:
  • Optimizing Expected Intersection-Over-Union With Candidate-Constrained CRFs [pdf] - Faruk Ahmed, Dany Tarlow, Dhruv Batra
  • Motion Trajectory Segmentation via Minimum Cost Multicuts [pdf] - Margret Keuper, Bjoern Andres, Thomas Brox
  • Efficient Video Segmentation Using Parametric Graph Partitioning [pdf] - Chen-Ping Yu, Hieu Le, Gregory Zelinsky, Dimitris Samaras
  • A Matrix Decomposition Perspective to Multiple Graph Matching [pdf] - Junchi Yan, Hongteng Xu, Hongyuan Zha, Xiaokang Yang, Huanxi Liu, Stephen Chu
  • Discrete Tabu Search for Graph Matching [pdf] - Kamil Adamczewski, Yumin Suh, Kyoung Mu Lee
Motion Trajectory Segmentation via Minimum Cost Multicuts [pdf]
Margret Keuper, Bjoern Andres, Thomas Brox

Abstract: For the segmentation of moving objects in videos, the analysis of long-term point trajectories has been very popular recently. In this paper, we formulate the segmentation of a video sequence based on point trajectories as a minimum cost multicut problem. Unlike the commonly used spectral clustering formulation, the minimum cost multicut formulation gives natural rise to optimize not only for a cluster assignment but also for the number of clusters while allowing for varying cluster sizes. In this setup, we provide a method to create a long-term point trajectory graph with attractive and repulsive binary terms and outperform state-of-the-art methods based on spectral clustering on the FBMS-59 dataset and on the motion subtask of the VSB100 dataset.
Similar papers:
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • Integrating Dashcam Views Through Inter-Video Mapping [pdf] - Hsin-I Chen, Yi-Ling Chen, Wei-Tse Lee, Fan Wang, Bing-Yu Chen
  • Action Detection by Implicit Intentional Motion Clustering [pdf] - Wei Chen, Jason J. Corso
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
  • Unsupervised Trajectory Clustering via Adaptive Multi-Kernel-Based Shrinkage [pdf] - Hongteng Xu, Yang Zhou, Weiyao Lin, Hongyuan Zha
Understanding and Predicting Image Memorability at a Large Scale [pdf]
Aditya Khosla, Akhil S. Raju, Antonio Torralba, Aude Oliva

Abstract: Progress in estimating visual memorability has been limited by the small scale and lack of variety of benchmark data. Here, we introduce a novel experimental procedure to objectively measure human memory, allowing us to build LaMem, the largest annotated image memorability dataset to date (containing 60,000 images from diverse sources). Using Convolutional Neural Networks (CNNs), we show that fine-tuned deep features outperform all other features by a large margin, reaching a rank correlation of 0.64, near human consistency (0.68). Analysis of the responses of the high-level CNN layers shows which objects and regions are positively, and negatively, correlated with memorability, allowing us to create memorability maps for each image and provide a concrete method to perform image memorability manipulation. This work demonstrates that one can now robustly estimate the memorability of images from many different classes, positioning memorability and deep memorability features as prime candidates to estimate the utility of information for cognitive systems. Our model and data are available at: http://memorability.csail.mit.edu
Similar papers:
  • Multi-Task Learning With Low Rank Attribute Embedding for Person Re-Identification [pdf] - Chi Su, Fan Yang, Shiliang Zhang, Qi Tian, Larry S. Davis, Wen Gao
  • SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf] - Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao
  • Improving Ferns Ensembles by Sparsifying and Quantising Posterior Probabilities [pdf] - Antonio L. Rodriguez, Vitor Sequeira
  • Learning Spatially Regularized Correlation Filters for Visual Tracking [pdf] - Martin Danelljan, Gustav Hager, Fahad Shahbaz Khan, Michael Felsberg
  • What Makes an Object Memorable? [pdf] - Rachit Dubey, Joshua Peterson, Aditya Khosla, Ming-Hsuan Yang, Bernard Ghanem
Where to Buy It: Matching Street Clothing Photos in Online Shops [pdf]
M. Hadi Kiapour, Xufeng Han, Svetlana Lazebnik, Alexander C. Berg, Tamara L. Berg

Abstract: In this paper, we define a new task, Exact Street to Shop, where our goal is to match a real-world example of a garment item to the same item in an online shop. This is an extremely challenging task due to visual differences between street photos (pictures of people wearing clothing in everyday uncontrolled settings) and online shop photos (pictures of clothing items on people, mannequins, or in isolation, captured by professionals in more controlled settings). We collect a new dataset for this application containing 404,683 shop photos collected from 25 different online retailers and 20,357 street photos, providing a total of 39,479 clothing item matches between street and shop photos. We develop three different methods for Exact Street to Shop retrieval, including two deep learning baseline methods, and a method to learn a similarity measure between the street and shop domains. Experiments demonstrate that our learned similarity significantly outperforms our baselines that use existing deep learning based representations.
Similar papers:
  • Multi-Label Cross-Modal Retrieval [pdf] - Viresh Ranjan, Nikhil Rasiwasia, C. V. Jawahar
  • Predicting Good Features for Image Geo-Localization Using Per-Bundle VLAD [pdf] - Hyo Jin Kim, Enrique Dunn, Jan-Michael Frahm
  • Localize Me Anywhere, Anytime: A Multi-Task Point-Retrieval Approach [pdf] - Guoyu Lu, Yan Yan, Li Ren, Jingkuan Song, Nicu Sebe, Chandra Kambhamettu
  • Learning Visual Clothing Style With Heterogeneous Dyadic Co-Occurrences [pdf] - Andreas Veit, Balazs Kovacs, Sean Bell, Julian McAuley, Kavita Bala, Serge Belongie
  • Cross-Domain Image Retrieval With a Dual Attribute-Aware Ranking Network [pdf] - Junshi Huang, Rogerio S. Feris, Qiang Chen, Shuicheng Yan
A Randomized Ensemble Approach to Industrial CT Segmentation [pdf]
Hyojin Kim, Jayaraman Jayaraman J. Thiagarajan, Peer-Timo Bremer

Abstract: Tuning the models and parameters of common segmentation approaches is challenging especially in the presence of noise and artifacts. Ensemble-based techniques attempt to compensate by randomly varying models and/or parameters to create a diverse set of hypotheses, which are subsequently ranked to arrive at the best solution. However, these methods have been restricted to cases where the underlying models are well established, e.g. natural images. In practice, it is difficult to determine a suitable base-model and the amount of randomization required. Furthermore, for multiobject scenes no single hypothesis may perform well for all objects, reducing the overall quality of the results. This paper presents a new ensemble-based segmentation framework for industrial CT images demonstrating that comparatively simple models and randomization strategies can significantly improve the result over existing techniques. Furthermore, we introduce a per-object based ranking, followed by a consensus inference that can outperform even the best case scenario of existing hypothesis ranking approaches. We demonstrate the effectiveness of our approach using a set of noise and artifact rich CT images from baggage security and show that it significantly outperforms existing solutions in this area.
Similar papers:
  • Learning Deconvolution Network for Semantic Segmentation [pdf] - Hyeonwoo Noh, Seunghoon Hong, Bohyung Han
  • An Accurate Iris Segmentation Framework Under Relaxed Imaging Constraints Using Total Variation Model [pdf] - Zijing Zhao, Kumar Ajay
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • Optimizing Expected Intersection-Over-Union With Candidate-Constrained CRFs [pdf] - Faruk Ahmed, Dany Tarlow, Dhruv Batra
Context-Guided Diffusion for Label Propagation on Graphs [pdf]
Kwang In Kim, James Tompkin, Hanspeter Pfister, Christian Theobalt

Abstract: Existing approaches for diffusion on graphs, e.g., for label propagation, are mainly focused on isotropic diffusion, which is induced by the commonly-used graph Laplacian regularizer. Inspired by the success of diffusivity tensors for anisotropic diffusion in image processing, we presents anisotropic diffusion on graphs and the corresponding label propagation algorithm. We develop positive definite diffusivity operators on the vector bundles of Riemannian manifolds, and discretize them to diffusivity operators on graphs. This enables us to easily define new robust diffusivity operators which significantly improve semi-supervised learning performance over existing diffusion algorithms.
Similar papers:
  • Pan-Sharpening With a Hyper-Laplacian Penalty [pdf] - Yiyong Jiang, Xinghao Ding, Delu Zeng, Yue Huang, John Paisley
  • A Wavefront Marching Method for Solving the Eikonal Equation on Cartesian Grids [pdf] - Brais Cancela, Marcos Ortega, Manuel G. Penedo
  • Multiple-Hypothesis Affine Region Estimation With Anisotropic LoG Filters [pdf] - Takahiro Hasegawa, Mitsuru Ambai, Kohta Ishikawa, Gou Koutaki, Yuji Yamauchi, Takayoshi Yamashita, Hironobu Fujiyoshi
  • Interpolation on the Manifold of K Component GMMs [pdf] - Hyunwoo J. Kim, Nagesh Adluru, Monami Banerjee, Baba C. Vemuri, Vikas Singh
  • Generic Promotion of Diffusion-Based Salient Object Detection [pdf] - Peng Jiang, Nuno Vasconcelos, Jingliang Peng
Interpolation on the Manifold of K Component GMMs [pdf]
Hyunwoo J. Kim, Nagesh Adluru, Monami Banerjee, Baba C. Vemuri, Vikas Singh

Abstract: Probability density functions (PDFs) are fundamental objects in mathematics with numerous applications in computer vision, machine learning and medical imaging. The feasibility of basic operations such as computing the distance between two PDFs and estimating a mean of a set of PDFs is a direct function of the representation we choose to work with. In this paper, we study the Gaussian mixture model (GMM) representation of the PDFs motivated by its numerous attractive features. (1) GMMs are arguably more interpretable than, say, square root parameterizations (2) the model complexity can be explicitly controlled by the number of components and (3) they are already widely used in many applications. The main contributions of this paper are numerical algorithms to enable basic operations on such objects that strictly respect their underlying geometry. For instance, when operating with a set of K component GMMs, a first order expectation is that the result of simple operations like interpolation and averaging should provide an object that is also a K component GMM. The literature provides very little guidance on enforcing such requirements systematically. It turns out that these tasks are important internal modules for analysis and processing of a field of ensemble average propagators (EAPs), common in diffusion weighted magnetic resonance imaging. We provide proof of principle experiments showing how the proposed algorithms for interpolation can facilitate statistical analysis of
Similar papers:
  • Beyond Gauss: Image-Set Matching on the Riemannian Manifold of PDFs [pdf] - Mehrtash Harandi, Mathieu Salzmann, Mahsa Baktashmotlagh
  • A Projection Free Method for Generalized Eigenvalue Problem With a Nonsmooth Regularizer [pdf] - Seong Jae Hwang, Maxwell D. Collins, Sathya N. Ravi, Vamsi K. Ithapu, Nagesh Adluru, Sterling C. Johnson, Vikas Singh
  • A Wavefront Marching Method for Solving the Eikonal Equation on Cartesian Grids [pdf] - Brais Cancela, Marcos Ortega, Manuel G. Penedo
  • Generic Promotion of Diffusion-Based Salient Object Detection [pdf] - Peng Jiang, Nuno Vasconcelos, Jingliang Peng
  • Context-Guided Diffusion for Label Propagation on Graphs [pdf] - Kwang In Kim, James Tompkin, Hanspeter Pfister, Christian Theobalt
Multiple Hypothesis Tracking Revisited [pdf]
Chanho Kim, Fuxin Li, Arridhana Ciptadi, James M. Rehg

Abstract: This paper revisits the classical multiple hypotheses tracking (MHT) algorithm in a tracking-by-detection framework. The success of MHT largely depends on the ability to maintain a small list of potential hypotheses, which can be facilitated with the accurate object detectors that are currently available. We demonstrate that a classical MHT implementation from the 90s can come surprisingly close to the performance of state-of-the-art methods on standard benchmark datasets. In order to further utilize the strength of MHT in exploiting higher-order information, we introduce a method for training online appearance models for each track hypothesis. We show that appearance models can be learned efficiently via a regularized least squares framework, requiring only a few extra operations for each hypothesis branch. We obtain state-of-the-art results on popular tracking-by-detection datasets such as PETS and the recent MOT challenge.
Similar papers:
  • A Versatile Learning-Based 3D Temporal Tracker: Scalable, Robust, Online [pdf] - David Joseph Tan, Federico Tombari, Slobodan Ilic, Nassir Navab
  • Learning to Track: Online Multi-Object Tracking by Decision Making [pdf] - Yu Xiang, Alexandre Alahi, Silvio Savarese
  • Near-Online Multi-Target Tracking With Aggregated Local Flow Descriptor [pdf] - Wongun Choi
  • Learning to Divide and Conquer for Online Multi-Target Tracking [pdf] - Francesco Solera, Simone Calderara, Rita Cucchiara
  • Joint Probabilistic Data Association Revisited [pdf] - Seyed Hamid Rezatofighi, Anton Milan, Zhen Zhang, Qinfeng Shi, Anthony Dick, Ian Reid
On Statistical Analysis of Neuroimages With Imperfect Registration [pdf]
Won Hwa Kim, Sathya N. Ravi, Sterling C. Johnson, Ozioma C. Okonkwo, Vikas Singh

Abstract: A variety of studies in neuroscience/neuroimaging seek to perform statistical inference on the acquired brain image scans for diagnosis as well as understanding the pathological manifestation of diseases. To do so, an important first step is to register (or co-register) all of the image data into a common coordinate system. This permits meaningful comparison of the intensities at each voxel across groups (e.g., diseased versus healthy) to evaluate the effects of the disease and/or use machine learning algorithms in a subsequent step. But errors in the underlying registration make this problematic, they either decrease the statistical power or make the follow-up inference tasks less effective/accurate. In this paper, we derive a novel algorithm which offers immunity to local errors in the underlying deformation field obtained from registration procedures. By deriving a deformation invariant representation of the image, the downstream analysis can be made more robust as if one had access to a (hypothetical) far superior registration procedure. Our algorithm is based on recent work on scattering transform. Using this as a starting point, we show how results from harmonic analysis (especially, nonEuclidean wavelets) yields strategies for designing deformation and additive noise invariant representations of large 3-D brain image volumes. We present a set of results on synthetic and real brain images where we achieve robust statistical analysis even in the presence of substantial d
Similar papers:
  • Pan-Sharpening With a Hyper-Laplacian Penalty [pdf] - Yiyong Jiang, Xinghao Ding, Delu Zeng, Yue Huang, John Paisley
  • A Groupwise Multilinear Correspondence Optimization for 3D Faces [pdf] - Timo Bolkart, Stefanie Wuhrer
  • An Adaptive Data Representation for Robust Point-Set Registration and Merging [pdf] - Dylan Campbell, Lars Petersson
  • Photometric Stereo in a Scattering Medium [pdf] - Zak Murez, Tali Treibitz, Ravi Ramamoorthi, David Kriegman
  • Airborne Three-Dimensional Cloud Tomography [pdf] - Aviad Levis, Yoav Y. Schechner, Amit Aides, Anthony B. Davis
Predicting Good Features for Image Geo-Localization Using Per-Bundle VLAD [pdf]
Hyo Jin Kim, Enrique Dunn, Jan-Michael Frahm

Abstract: We address the problem of recognizing a place depicted in a query image by using a large database of geo-tagged images at a city-scale. In particular, we discover features that are useful for recognizing a place in a data-driven manner, and use this knowledge to predict useful features in a query image prior to the geo-localization process. This allows us to achieve better performance while reducing the number of features. Also, for both learning to predict features and retrieving geo-tagged images from the database, we propose per-bundle vector of locally aggregated descriptors (PBVLAD), where each maximally stable region is described by a vector of locally aggregated descriptors (VLAD) on multiple scale-invariant features detected within the region. Experimental results show the proposed approach achieves a significant improvement over other baseline methods.
Similar papers:
  • Query Adaptive Similarity Measure for RGB-D Object Recognition [pdf] - Yanhua Cheng, Rui Cai, Chi Zhang, Zhiwei Li, Xin Zhao, Kaiqi Huang, Yong Rui
  • Attribute-Graph: A Graph Based Approach to Image Ranking [pdf] - Nikita Prabhu, R. Venkatesh Babu
  • RIDE: Reversal Invariant Descriptor Enhancement [pdf] - Lingxi Xie, Jingdong Wang, Weiyao Lin, Bo Zhang, Qi Tian
  • Localize Me Anywhere, Anytime: A Multi-Task Point-Retrieval Approach [pdf] - Guoyu Lu, Yan Yan, Li Ren, Jingkuan Song, Nicu Sebe, Chandra Kambhamettu
  • Hyperpoints and Fine Vocabularies for Large-Scale Location Recognition [pdf] - Torsten Sattler, Michal Havlena, Filip Radenovic, Konrad Schindler, Marc Pollefeys
Simultaneous Foreground Detection and Classification With Hybrid Features [pdf]
Jaemyun Kim, Adin Ramirez Rivera, Byungyong Ryu, Oksam Chae

Abstract: ... In this paper, we propose a hybrid background model that relies on edge and non-edge features of the image to produce the model. We encode these features into a coding scheme, that we called Local Hybrid Pattern (LHP), that selectively models edges and non-edges features of each pixel. Furthermore, we model each pixel with an adaptive code dictionary to represent the background dynamism, and update it by adding stable codes and discarding unstable ones. We weight each code in the dictionary to enhance its description of the pixel it models. The foreground is detected as the incoming codes that deviate from the dictionary. We can detect (as foreground or background) and classify (as edge or inner region) each pixel simultaneously. We tested our proposed method in existing databases with promising results.
Similar papers:
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
  • Holistically-Nested Edge Detection [pdf] - Saining Xie, Zhuowen Tu
  • Two Birds, One Stone: Jointly Learning Binary Code for Large-Scale Face Image Retrieval and Attributes Prediction [pdf] - Yan Li, Ruiping Wang, Haomiao Liu, Huajie Jiang, Shiguang Shan, Xilin Chen
  • Learning Binary Codes for Maximum Inner Product Search [pdf] - Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen
  • Image Matting With KL-Divergence Based Sparse Sampling [pdf] - Levent Karacan, Aykut Erdem, Erkut Erdem
SOWP: Spatially Ordered and Weighted Patch Descriptor for Visual Tracking [pdf]
Han-Ul Kim, Dae-Youn Lee, Jae-Young Sim, Chang-Su Kim

Abstract: A simple yet effective object descriptor for visual tracking is proposed in this paper. We first decompose the bounding box of a target object into multiple patches, which are described by color and gradient histograms. Then, we concatenate the features of the spatially ordered patches to represent the object appearance. Moreover, to alleviate the impacts of background information possibly included in the bounding box, we determine patch weights using random walk with restart (RWR) simulations. The patch weights represent the importance of each patch in the description of foreground information, and are used to construct an object descriptor, called spatially ordered and weighted patch (SOWP) descriptor. We incorporate the proposed SOWP descriptor into the structured output tracking framework. Experimental results demonstrate that the proposed algorithm yields significantly better performance than the state-of-theart trackers on a recent benchmark dataset, and also excels in another recent benchmark dataset.
Similar papers:
  • Visual Tracking With Fully Convolutional Networks [pdf] - Lijun Wang, Wanli Ouyang, Xiaogang Wang, Huchuan Lu
  • Discriminative Low-Rank Tracking [pdf] - Yao Sui, Yafei Tang, Li Zhang
  • Online Object Tracking With Proposal Selection [pdf] - Yang Hua, Karteek Alahari, Cordelia Schmid
  • Tracking-by-Segmentation With Online Gradient Boosting Decision Tree [pdf] - Jeany Son, Ilchae Jung, Kayoung Park, Bohyung Han
  • Learning Spatially Regularized Correlation Filters for Visual Tracking [pdf] - Martin Danelljan, Gustav Hager, Fahad Shahbaz Khan, Michael Felsberg
Inferring M-Best Diverse Labelings in a Single One [pdf]
Alexander Kirillov, Bogdan Savchynskyy, Dmitrij Schlesinger, Dmitry Vetrov, Carsten Rother

Abstract: We consider the task of finding M -best diverse solutions in a graphical model. In a previous work by Batra et al. an algorithmic approach for finding such solutions was proposed, and its usefulness was shown in numerous applications. Contrary to previous work we propose a novel formulation of the problem in form of a single energy minimization problem in a specially constructed graphical model. We show that the method of Batra et al. can be considered as a greedy approximate algorithm for our model, whereas we introduce an efficient specialized optimization technique for it, based on alpha-expansion. We evaluate our method on two application scenarios, interactive and semantic image segmentation, with binary and multiple labels. In both cases we achieve considerably better error rates than stateof-the art diversity methods. Furthermore, we empirically discover that in the binary label case we were able to reach global optimality for all test instances.
Similar papers:
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
  • A Multiscale Variable-Grouping Framework for MRF Energy Minimization [pdf] - Omer Meir, Meirav Galun, Stav Yagev, Ronen Basri, Irad Yavneh
  • Predicting Multiple Structured Visual Interpretations [pdf] - Debadeepta Dey, Varun Ramakrishna, Martial Hebert, J. Andrew Bagnell
  • Optimizing Expected Intersection-Over-Union With Candidate-Constrained CRFs [pdf] - Faruk Ahmed, Dany Tarlow, Dhruv Batra
  • Parsimonious Labeling [pdf] - Puneet K. Dokania, M. Pawan Kumar
Unsupervised Domain Adaptation for Zero-Shot Learning [pdf]
Elyor Kodirov, Tao Xiang, Zhenyong Fu, Shaogang Gong

Abstract: Zero-shot learning (ZSL) can be considered as a special case of transfer learning where the source and target domains have different tasks/label spaces and the target domain is unlabelled, providing little guidance for the knowledge transfer. A ZSL method typically assumes that the two domains share a common semantic representation space, where a visual feature vector extracted from an image/video can be projected/embedded using a projection function. Existing approaches learn the projection function from the source domain and apply it without adaptation to the target domain. They are thus based on naive knowledge transfer and the learned projections are prone to the domain shift problem. In this paper a novel ZSL method is proposed based on unsupervised domain adaptation. Specically, we formulate a novel regularised sparse coding framework which uses the target domain class labels projections in the semantic space to regularise the learned target domain projection thus effectively overcoming the projection domain shift problem. Extensive experiments on four object and action recognition benchmark datasets show that the proposed ZSL method signicantly outperforms the state-of-the-arts. L tiger bear zebra whale Label Embedding L Label Space A Attribute Space F Feature Space Prototypes A F A Feature Embedding F Source Domain Target Domain
Similar papers:
  • Multi-View Domain Generalization for Visual Recognition [pdf] - Li Niu, Wen Li, Dong Xu
  • Simultaneous Deep Transfer Across Domains and Tasks [pdf] - Eric Tzeng, Judy Hoffman, Trevor Darrell, Kate Saenko
  • Unsupervised Domain Adaptation With Imbalanced Cross-Domain Data [pdf] - Tzu Ming Harry Hsu, Wei Yu Chen, Cheng-An Hou, Yao-Hung Hubert Tsai, Yi-Ren Yeh, Yu-Chiang Frank Wang
  • Bi-Shifting Auto-Encoder for Unsupervised Domain Adaptation [pdf] - Meina Kan, Shiguang Shan, Xilin Chen
  • Zero-Shot Learning via Semantic Similarity Embedding [pdf] - Ziming Zhang, Venkatesh Saligrama
Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf]
Naejin Kong, Michael J. Black

Abstract: We formulate the estimation of dense depth maps from video sequences as a problem of intrinsic image estimation. Our approach synergistically integrates the estimation of multiple intrinsic images including depth, albedo, shading, optical flow, and surface contours. We build upon an example-based framework for depth estimation that uses label transfer from a database of RGB and depth pairs. We combine this with a method that extracts consistent albedo and shading from video. In contrast to raw RGB values, albedo and shading provide a richer, more physical, foundation for depth transfer. Additionally we train a new contour detector to predict surface boundaries from albedo, shading, and pixel values and use this to improve the estimation of depth boundaries. We also integrate sparse structure from motion with our method to improve the metric accuracy of the estimated depth maps. We evaluate our Intrinsic Depth method quantitatively by estimating depth from videos in the NYU RGB-D and SUN3D datasets. We find that combining the estimation of multiple intrinsic images improves depth estimation relative to the baseline method.
Similar papers:
  • Polarized 3D: High-Quality Depth Sensing With Polarization Cues [pdf] - Achuta Kadambi, Vage Taamazyan, Boxin Shi, Ramesh Raskar
  • High Quality Structure From Small Motion for Rolling Shutter Cameras [pdf] - Sunghoon Im, Hyowon Ha, Gyeongmin Choe, Hae-Gon Jeon, Kyungdon Joo, In So Kweon
  • Learning Ordinal Relationships for Mid-Level Vision [pdf] - Daniel Zoran, Phillip Isola, Dilip Krishnan, William T. Freeman
  • Direct Intrinsics: Learning Albedo-Shading Decomposition by Convolutional Regression [pdf] - Takuya Narihira, Michael Maire, Stella X. Yu
  • Intrinsic Scene Decomposition From RGB-D images [pdf] - Mohammed Hachama, Bernard Ghanem, Peter Wonka
Deep Neural Decision Forests [pdf]
Peter Kontschieder, Madalina Fiterau, Antonio Criminisi, Samuel Rota Bulo

Abstract: We present Deep Neural Decision Forests a novel approach that unifies classification trees with the representation learning functionality known from deep convolutional networks, by training them in an end-to-end manner. To combine these two worlds, we introduce a stochastic and differentiable decision tree model, which steers the representation learning usually conducted in the initial layers of a (deep) convolutional network. Our model differs from conventional deep networks because a decision forest provides the final predictions and it differs from conventional decision forests since we propose a principled, joint and global optimization of split and leaf node parameters. We show experimental results on benchmark machine learning datasets like MNIST and ImageNet and find onpar or superior results when compared to state-of-the-art deep models. Most remarkably, we obtain Top5-Errors of only 7.84%/6.38% on ImageNet validation data when integrating our forests in a single-crop, single/seven model GoogLeNet architecture, respectively. Thus, even without any form of training data set augmentation we are improving on the 6.67% error obtained by the best GoogLeNet architecture (7 models, 144 crops).
Similar papers:
  • COUNT Forest: CO-Voting Uncertain Number of Targets Using Random Forest for Crowd Density Estimation [pdf] - Viet-Quoc Pham, Tatsuo Kozakaya, Osamu Yamaguchi, Ryuzo Okada
  • Pairwise Conditional Random Forests for Facial Expression Recognition [pdf] - Arnaud Dapogny, Kevin Bailly, Severine Dubuisson
  • Tracking-by-Segmentation With Online Gradient Boosting Decision Tree [pdf] - Jeany Son, Ilchae Jung, Kayoung Park, Bohyung Han
  • Naive Bayes Super-Resolution Forest [pdf] - Jordi Salvador, Eduardo Perez-Pellitero
  • Conditional Convolutional Neural Network for Modality-Aware Face Recognition [pdf] - Chao Xiong, Xiaowei Zhao, Danhang Tang, Karlekar Jayashree, Shuicheng Yan, Tae-Kyun Kim
Introducing Geometry in Active Learning for Image Segmentation [pdf]
Ksenia Konyushkova, Raphael Sznitman, Pascal Fua

Abstract: We propose an Active Learning approach to training a segmentation classifier that exploits geometric priors to streamline the annotation process in 3D image volumes. To this end, we use these priors not only to select voxels most in need of annotation but to guarantee that they lie on 2D planar patch, which makes it much easier to annotate than if they were randomly distributed in the volume. A simplified version of this approach is effective in natural 2D images. We evaluated our approach on Electron Microscopy and Magnetic Resonance image volumes, as well as on natural images. Comparing our approach against several accepted baselines demonstrates a marked performance increase. larly challenging, in part because it is difficult both to develop effective interfaces to visualize the huge image data and for users to quickly figure out what they are looking at. In this paper, we will therefore focus on image volumes but the techniques we will discuss are nevertheless also applicable to regular 2D images by treating them as stacks of height one. With this, we introduce here a novel approach to AL that is geared towards segmenting 3D image volumes and also applicable to ordinary 2D images. By design, it takes into account geometric constraints to which regions should obey and makes the annotation process convenient. Our contribution hence is twofold: We introduce a way to exploit geometric priors to more effectively select the image data the expert user is asked to annotate.
Similar papers:
  • Entropy-Based Latent Structured Output Prediction [pdf] - Diane Bouchacourt, Sebastian Nowozin, M. Pawan Kumar
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
  • Efficient Classifier Training to Minimize False Merges in Electron Microscopy Segmentation [pdf] - Toufiq Parag, Dan C. Ciresan, Alessandro Giusti
  • Multi-Class Multi-Annotator Active Learning With Robust Gaussian Process for Visual Recognition [pdf] - Chengjiang Long, Gang Hua
  • Active Transfer Learning With Zero-Shot Priors: Reusing Past Datasets for Future Tasks [pdf] - Efstratios Gavves, Thomas Mensink, Tatiana Tommasi, Cees G. M. Snoek, Tinne Tuytelaars
Peeking Template Matching for Depth Extension [pdf]
Simon Korman, Eyal Ofek, Shai Avidan

Abstract: We propose a method that extends a given depth image into regions in 3D that are not visible from the point of view of the camera. The algorithm detects repeated 3D structures in the visible scene and suggests a set of 3D extension hypotheses, which are then combined together through a global 3D MRF discrete optimization. The recovered global 3D surface is consistent with both the input depth map and the hypotheses. A key component of this work is a novel 3D template matcher that is used to detect repeated 3D structure in the scene and to suggest the hypotheses. A unique property of this matcher is that it can handle depth uncertainty. This is crucial because the matcher is required to peek around the corner, as it operates at the boundaries of the visible 3D scene where depth information is missing. The proposed matcher is fast and is guaranteed to find an approximation to the globally optimal solution. We demonstrate on real-world data that our algorithm is capable of completing a full 3D scene from a single depth image and can synthesize a full depth map from a novel viewpoint of the scene. In addition, we report results on an extensive synthetic set of 3D shapes, which allows us to evaluate the method both qualitatively and quantitatively.
Similar papers:
  • Exploiting Object Similarity in 3D Reconstruction [pdf] - Chen Zhou, Fatma Guney, Yizhou Wang, Andreas Geiger
  • As-Rigid-As-Possible Volumetric Shape-From-Template [pdf] - Shaifali Parashar, Daniel Pizarro, Adrien Bartoli, Toby Collins
  • Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf] - Naejin Kong, Michael J. Black
  • Dense Image Registration and Deformable Surface Reconstruction in Presence of Occlusions and Minimal Texture [pdf] - Dat Tien Ngo, Sanghyuk Park, Anne Jorstad, Alberto Crivellaro, Chang D. Yoo, Pascal Fua
  • Polarized 3D: High-Quality Depth Sensing With Polarization Cues [pdf] - Achuta Kadambi, Vage Taamazyan, Boxin Shi, Ramesh Raskar
Probabilistic Appearance Models for Segmentation and Classification [pdf]
Julia Kruger, Jan Ehrhardt, Heinz Handels

Abstract: Statistical shape and appearance models are often based on the accurate identification of one-to-one correspondences in a training data set. At the same time, the determination of these corresponding landmarks is the most challenging part of such methods. Hufnagel et al. developed an alternative method using correspondence probabilities for a statistical shape model. We propose the use of probabilistic correspondences for statistical appearance models by incorporating appearance information into the framework. A point-based representation is employed representing the image by a set of vectors assembling position and appearances. Using probabilistic correspondences between these multi-dimensional feature vectors eliminates the need for extensive preprocessing to find corresponding landmarks and reduces the dependence of the generated model on the landmark positions. Then, a maximum a-posteriori approach is used to derive a single global optimization criterion with respect to model parameters and observation dependent parameters, that directly affects shape and appearance information of the considered structures. Model generation and fitting can be expressed by optimizing the same criterion. The developed framework describes the modeling process in a concise and flexible mathematical way and allows for additional constraints as topological regularity in the modeling process. Furthermore, it eliminates the demand for costly correspondence determination. We apply the model for se
Similar papers:
  • Compositional Hierarchical Representation of Shape Manifolds for Classification of Non-Manifold Shapes [pdf] - Mete Ozay, Umit Rusen Aktas, Jeremy L. Wyatt, Ales Leonardis
  • Person Re-Identification With Correspondence Structure Learning [pdf] - Yang Shen, Weiyao Lin, Junchi Yan, Mingliang Xu, Jianxin Wu, Jingdong Wang
  • Contour Flow: Middle-Level Motion Estimation by Combining Motion Segmentation and Contour Alignment [pdf] - Huijun Di, Qingxuan Shi, Feng Lv, Ming Qin, Yao Lu
  • Single Image Pop-Up From Discriminatively Learned Parts [pdf] - Menglong Zhu, Xiaowei Zhou, Kostas Daniilidis
  • The One Triangle Three Parallelograms Sampling Strategy and Its Application in Shape Regression [pdf] - Mikael Nilsson
Learning Analysis-by-Synthesis for 6D Pose Estimation in RGB-D Images [pdf]
Alexander Krull, Eric Brachmann, Frank Michel, Michael Ying Yang, Stefan Gumhold, Carsten Rother

Abstract: Analysis-by-synthesis has been a successful approach for many tasks in computer vision, such as 6D pose estimation of an object in an RGB-D image which is the topic of this work. The idea is to compare the observation with the output of a forward process, such as a rendered image of the object of interest in a particular pose. Due to occlusion or complicated sensor noise, it can be difficult to perform this comparison in a meaningful way. We propose an approach that learns to compare, while taking these difficulties into account. This is done by describing the posterior density of a particular object pose with a convolutional neural network (CNN) that compares observed and rendered images. The network is trained with the maximum likelihood paradigm. We observe empirically that the CNN does not specialize to the geometry or appearance of specific objects. It can be used with objects of vastly different shapes and appearances, and in different backgrounds. Compared to state-of-the-art, we demonstrate a significant improvement on two different datasets which include a total of eleven objects, cluttered background, and heavy occlusion.
Similar papers:
  • Semantic Pose Using Deep Networks Trained on Synthetic RGB-D [pdf] - Jeremie Papon, Markus Schoeler
  • Training a Feedback Loop for Hand Pose Estimation [pdf] - Markus Oberweger, Paul Wohlhart, Vincent Lepetit
  • Pose Induction for Novel Object Categories [pdf] - Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • Opening the Black Box: Hierarchical Sampling Optimization for Estimating Human Hand Pose [pdf] - Danhang Tang, Jonathan Taylor, Pushmeet Kohli, Cem Keskin, Tae-Kyun Kim, Jamie Shotton
Efficient Solution to the Epipolar Geometry for Radially Distorted Cameras [pdf]
Zuzana Kukelova, Jan Heller, Martin Bujnak, Andrew Fitzgibbon, Tomas Pajdla

Abstract: The estimation of the epipolar geometry of two cameras from image matches is a fundamental problem of computer vision with many applications. While the closely related problem of estimating relative pose of two different uncalibrated cameras with radial distortion is of particular importance, none of the previously published methods is suitable for practical applications. These solutions are either numerically unstable, sensitive to noise, based on a large number of point correspondences, or simply too slow for real-time applications. In this paper, we present a new efficient solution to this problem that uses 10 image correspondences. By manipulating ten input polynomial equations, we derive a degree 10 polynomial equation in one variable. The solutions to this equation are efficiently found using the Sturm sequences method. In the experiments, we show that the proposed solution is stable, noise resistant, and fast, and as such efficiently usable in a practical Structure-fromMotion pipeline.
Similar papers:
  • Minimal Solvers for 3D Geometry From Satellite Imagery [pdf] - Enliang Zheng, Ke Wang, Enrique Dunn, Jan-Michael Frahm
  • Non-Parametric Structure-Based Calibration of Radially Symmetric Cameras [pdf] - Federico Camposeco, Torsten Sattler, Marc Pollefeys
  • An Efficient Minimal Solution for Multi-Camera Motion [pdf] - Jonathan Ventura, Clemens Arth, Vincent Lepetit
  • On the Equivalence of Moving Entrance Pupil and Radial Distortion for Camera Calibration [pdf] - Avinash Kumar, Narendra Ahuja
  • Structure From Motion Using Structure-Less Resection [pdf] - Enliang Zheng, Changchang Wu
On the Equivalence of Moving Entrance Pupil and Radial Distortion for Camera Calibration [pdf]
Avinash Kumar, Narendra Ahuja

Abstract: Radial distortion for ordinary (non-fisheye) camera lenses has traditionally been modeled as an infinite series function of radial location of an image pixel from the image center. While there has been enough empirical evidence to show that such a model is accurate and sufficient for radial distortion calibration, there has not been much analysis on the geometric/physical understanding of radial distortion from a camera calibration perspective. In this paper, we show using a thick-lens imaging model, that the variation of entrance pupil location as a function of incident image ray angle is directly responsible for radial distortion in captured images. Thus, unlike as proposed in the current state-of-the-art in camera calibration, radial distortion and entrance pupil movement are equivalent and need not be modeled together. By modeling only entrance pupil motion instead of radial distortion, we achieve two main benefits; first, we obtain comparable if not better pixel re-projection error than traditional methods; second, and more importantly, we directly back-project a radially distorted image pixel along the true image ray which formed it. Using a thick-lens setting, we show that such a back-projection is more accurate than the two-step method of undistorting an image pixel and then back-projecting it. We have applied this calibration method to the problem of generative depth-from-focus using focal stack to get accurate depth estimates. dinary (non-fisheye) camera lenses who
Similar papers:
  • Extended Depth of Field Catadioptric Imaging Using Focal Sweep [pdf] - Ryunosuke Yokoya, Shree K. Nayar
  • Accurate Camera Calibration Robust to Defocus Using a Smartphone [pdf] - Hyowon Ha, Yunsu Bok, Kyungdon Joo, Jiyoung Jung, In So Kweon
  • A Linear Generalized Camera Calibration From Three Intersecting Reference Planes [pdf] - Mai Nishimura, Shohei Nobuhara, Takashi Matsuyama, Shinya Shimizu, Kensaku Fujii
  • Efficient Solution to the Epipolar Geometry for Radially Distorted Cameras [pdf] - Zuzana Kukelova, Jan Heller, Martin Bujnak, Andrew Fitzgibbon, Tomas Pajdla
  • Non-Parametric Structure-Based Calibration of Radially Symmetric Cameras [pdf] - Federico Camposeco, Torsten Sattler, Marc Pollefeys
Visual Phrases for Exemplar Face Detection [pdf]
Vijay Kumar, Anoop Namboodiri, C. V. Jawahar

Abstract: Recently, exemplar based approaches [13, 22] have been successfully applied for face detection in the wild. Contrary to traditional approaches that model face variations from a large and diverse set of training examples, exemplar-based approaches use a collection of discriminatively trained exemplars for detection. In this paradigm, each exemplar casts a vote using retrieval framework and generalized Hough voting, to locate the faces in the target image. The advantage of this approach is that by having a large database that covers all possible variations, faces in challenging conditions can be detected without having to learn explicit models for different variations. Current schemes, however, make an assumption of independence between the visual words, ignoring their relations in the process. They also ignore the spatial consistency of the visual words. Consequently, every exemplar word contributes equally during voting regardless of its location. In this paper, we propose a novel approach that incorporates higher order information in the voting process. We discover visual phrases that contain semantically related visual words and exploit them for detection along with the visual words. For spatial consistency, we estimate the spatial distribution of visual words and phrases from the entire database and then weigh their occurrence in exemplars. This ensures that a visual word or a phrase in an exemplar makes a major contribution only if it occurs at its semantic location, ther
Similar papers:
  • Multi-View Complementary Hash Tables for Nearest Neighbor Search [pdf] - Xianglong Liu, Lei Huang, Cheng Deng, Jiwen Lu, Bo Lang
  • Multi-View Domain Generalization for Visual Recognition [pdf] - Li Niu, Wen Li, Dong Xu
  • Flickr30k Entities: Collecting Region-to-Phrase Correspondences for Richer Image-to-Sentence Models [pdf] - Bryan A. Plummer, Liwei Wang, Chris M. Cervantes, Juan C. Caicedo, Julia Hockenmaier, Svetlana Lazebnik
  • Segment-Phrase Table for Semantic Segmentation, Visual Entailment and Paraphrasing [pdf] - Hamid Izadinia, Fereshteh Sadeghi, Santosh K. Divvala, Hannaneh Hajishirzi, Yejin Choi, Ali Farhadi
  • Common Subspace for Model and Similarity: Phrase Learning for Caption Generation From Images [pdf] - Yoshitaka Ushiku, Masataka Yamaguchi, Yusuke Mukuta, Tatsuya Harada
DeepBox: Learning Objectness With Convolutional Networks [pdf]
Weicheng Kuo, Bharath Hariharan, Jitendra Malik

Abstract: Existing object proposal approaches use primarily bottom-up cues to rank proposals, while we believe that objectness is in fact a high level construct. We argue for a data-driven, semantic approach for ranking object proposals. Our framework, which we call DeepBox, uses convolutional neural networks (CNNs) to rerank proposals from a bottom-up method. We use a novel four-layer CNN architecture that is as good as much larger networks on the task of evaluating objectness while being much faster. We show that DeepBox significantly improves over the bottomup ranking, achieving the same recall with 500 proposals as achieved by bottom-up methods with 2000. This improvement generalizes to categories the CNN has never seen before and leads to a 4.5-point gain in detection mAP. Our implementation achieves this performance while running at 260 ms per image.
Similar papers:
  • Boosting Object Proposals: From Pascal to COCO [pdf] - Jordi Pont-Tuset, Luc Van Gool
  • Box Aggregation for Proposal Decimation: Last Mile of Object Detection [pdf] - Shu Liu, Cewu Lu, Jiaya Jia
  • Oriented Object Proposals [pdf] - Shengfeng He, Rynson W.H. Lau
  • DeepProposal: Hunting Objects by Cascading Deep Convolutional Layers [pdf] - Amir Ghodrati, Ali Diba, Marco Pedersoli, Tinne Tuytelaars, Luc Van Gool
  • Cascaded Sparse Spatial Bins for Efficient and Effective Generic Object Detection [pdf] - David Novotny, Jiri Matas
Unsupervised Object Discovery and Tracking in Video Collections [pdf]
Suha Kwak, Minsu Cho, Ivan Laptev, Jean Ponce, Cordelia Schmid

Abstract: This paper addresses the problem of automatically localizing dominant objects as spatio-temporal tubes in a noisy collection of videos with minimal or even no supervision. We formulate the problem as a combination of two complementary processes: discovery and tracking. The first one establishes correspondences between prominent regions across videos, and the second one associates similar object regions within the same video. Interestingly, our algorithm also discovers the implicit topology of frames associated with instances of the same object class across different videos, a role normally left to supervisory information in the form of class labels in conventional image and video understanding methods. Indeed, as demonstrated by our experiments, our method can handle video collections featuring multiple object classes, and substantially outperforms the state of the art in colocalization, even though it tackles a broader problem with much less supervision.
Similar papers:
  • Self-Occlusions and Disocclusions in Causal Video Object Segmentation [pdf] - Yanchao Yang, Ganesh Sundaramoorthi, Stefano Soatto
  • Fully Connected Object Proposals for Video Segmentation [pdf] - Federico Perazzi, Oliver Wang, Markus Gross, Alexander Sorkine-Hornung
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
  • Learning to Track for Spatio-Temporal Action Localization [pdf] - Philippe Weinzaepfel, Zaid Harchaoui, Cordelia Schmid
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
RGB-Guided Hyperspectral Image Upsampling [pdf]
Hyeokhyen Kwon, Yu-Wing Tai

Abstract: Hyperspectral imaging usually lack of spatial resolution due to limitations of hardware design of imaging sensors. On the contrary, latest imaging sensors capture a RGB image with resolution of multiple times larger than a hyperspectral image. In this paper, we present an algorithm to enhance and upsample the resolution of hyperspectral images. Our algorithm consists of two stages: spatial upsampling stage and spectrum substitution stage. The spatial upsampling stage is guided by a high resolution RGB image of the same scene, and the spectrum substitution stage utilizes sparse coding to locally refine the upsampled hyperspectral image through dictionary substitution. Experiments show that our algorithm is highly effective and has outperformed state-of-the-art matrix factorization based approaches.
Similar papers:
  • Variational Depth Superresolution Using Example-Based Edge Representations [pdf] - David Ferstl, Matthias Ruther, Horst Bischof
  • Hyperspectral Compressive Sensing Using Manifold-Structured Sparsity Prior [pdf] - Lei Zhang, Wei Wei, Yanning Zhang, Fei Li, Chunhua Shen, Qinfeng Shi
  • Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf] - Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • Separating Fluorescent and Reflective Components by Using a Single Hyperspectral Image [pdf] - Yinqiang Zheng, Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • Hyperspectral Super-Resolution by Coupled Spectral Unmixing [pdf] - Charis Lanaras, Emmanuel Baltsavias, Konrad Schindler
Intrinsic Decomposition of Image Sequences From Local Temporal Variations [pdf]
Pierre-Yves Laffont, Jean-Charles Bazin

Abstract: We present a method for intrinsic image decomposition, which aims to decompose images into reflectance and shading layers. Our input is a sequence of images with varying illumination acquired by a static camera, e.g. an indoor scene with a moving light source or an outdoor timelapse. We leverage the local color variations observed over time to infer constraints on the reflectance and solve the ill-posed image decomposition problem. In particular, we derive an adaptive local energy from the observations of each local neighborhood over time, and integrate distant pairwise constraints to enforce coherent decomposition across all surfaces with consistent shading changes. Our method is solely based on multiple observations of a Lambertian scene under varying illumination and does not require user interaction, scene geometry, or an explicit lighting model. We compare our results with several intrinsic decomposition methods on a number of synthetic and captured datasets. Jean-Charles Bazin ETH Zurich age decomposition, where the input is an image sequence of a static scene with fixed viewpoint under varying illumination. We show that the observation of temporal variations in local neighborhoods provides us with all the necessary information to relate the reflectance of multiple pixels, both at the local level and across pairs of distant pixels. This allows us to solve the decomposition problem and obtain intrinsic layers without the need of geometry, user interaction, an explicit
Similar papers:
  • Direct Intrinsics: Learning Albedo-Shading Decomposition by Convolutional Regression [pdf] - Takuya Narihira, Michael Maire, Stella X. Yu
  • Intrinsic Scene Decomposition From RGB-D images [pdf] - Mohammed Hachama, Bernard Ghanem, Peter Wonka
  • Learning Ordinal Relationships for Mid-Level Vision [pdf] - Daniel Zoran, Phillip Isola, Dilip Krishnan, William T. Freeman
  • A Comprehensive Multi-Illuminant Dataset for Benchmarking of the Intrinsic Image Algorithms [pdf] - Shida Beigpour, Andreas Kolb, Sven Kunz
  • Learning Data-Driven Reflectance Priors for Intrinsic Image Decomposition [pdf] - Tinghui Zhou, Philipp Krahenbuhl, Alexei A. Efros
Robust Heart Rate Measurement From Video Using Select Random Patches [pdf]
Antony Lam, Yoshinori Kuno

Abstract: The ability to remotely measure heart rate from videos without requiring any special setup is beneficial to many applications. In recent years, a number of papers on heart rate (HR) measurement from videos have been proposed. However, these methods typically require the human subject to be stationary and for the illumination to be controlled. For methods that do take into account motion and illumination changes, strong assumptions are still made about the environment (e.g. background can be used for illumination rectification). In this paper, we propose an HR measurement method that is robust to motion, illumination changes, and does not require use of an environments background. We present conditions under which cardiac activity extraction from local regions of the face can be treated as a linear Blind Source Separation problem and propose a simple but robust algorithm for selecting good local regions. The independent HR estimates from multiple local regions are then combined in a majority voting scheme that robustly recovers the HR. We validate our algorithm on a large database of challenging videos. (a) Proposed Video Heart Rate Measurement System (b) Estimated Plethysmographic Signal from Majority Vote vs. Ground Truth ECG Figure 1: Heart Rate Measurement from Video Captured by Consumer Camera mentary information in a facial emotion recognition system, which would be of great benefit to affective computing in human-computer and human-robot interaction applications. An
Similar papers:
  • FaceDirector: Continuous Control of Facial Performance in Video [pdf] - Charles Malleson, Jean-Charles Bazin, Oliver Wang, Derek Bradley, Thabo Beeler, Adrian Hilton, Alexander Sorkine-Hornung
  • A Comprehensive Multi-Illuminant Dataset for Benchmarking of the Intrinsic Image Algorithms [pdf] - Shida Beigpour, Andreas Kolb, Sven Kunz
  • Nighttime Haze Removal With Glow and Multiple Light Colors [pdf] - Yu Li, Robby T. Tan, Michael S. Brown
  • Video Restoration Against Yin-Yang Phasing [pdf] - Xiaolin Wu, Zhenhao Li, Xiaowei Deng
  • Separating Fluorescent and Reflective Components by Using a Single Hyperspectral Image [pdf] - Yinqiang Zheng, Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
Action Recognition by Hierarchical Mid-Level Action Elements [pdf]
Tian Lan, Yuke Zhu, Amir Roshan Zamir, Silvio Savarese

Abstract: Realistic videos of human actions exhibit rich spatiotemporal structures at multiple levels of granularity: an action can always be decomposed into multiple finer-grained elements in both space and time. To capture this intuition, we propose to represent videos by a hierarchy of mid-level action elements (MAEs), where each MAE corresponds to an action-related spatiotemporal segment in the video. We introduce an unsupervised method to generate this representation from videos. Our method is capable of distinguishing action-related segments from background segments and representing actions at multiple spatiotemporal resolutions. Given a set of spatiotemporal segments generated from the training data, we introduce a discriminative clustering algorithm that automatically discovers MAEs at multiple levels of granularity. We develop a structured model that captures a rich set of spatial, temporal and hierarchical relations among the segments, where the action label and multiple levels of MAE labels are jointly inferred. The proposed model achieves state-of-the-art performance in multiple action recognition benchmarks. Moreover, we demonstrate the effectiveness of our model in real-world applications such as action recognition in large-scale untrimmed videos and action parsing.
Similar papers:
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
  • Action Detection by Implicit Intentional Motion Clustering [pdf] - Wei Chen, Jason J. Corso
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Action Localization in Videos Through Context Walk [pdf] - Khurram Soomro, Haroon Idrees, Mubarak Shah
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
Hyperspectral Super-Resolution by Coupled Spectral Unmixing [pdf]
Charis Lanaras, Emmanuel Baltsavias, Konrad Schindler

Abstract: Hyperspectral cameras capture images with many narrow spectral channels, which densely sample the electromagnetic spectrum. The detailed spectral resolution is useful for many image analysis problems, but it comes at the cost of much lower spatial resolution. Hyperspectral super-resolution addresses this problem, by fusing a lowresolution hyperspectral image and a conventional highresolution image into a product of both high spatial and high spectral resolution. In this paper, we propose a method which performs hyperspectral super-resolution by jointly unmixing the two input images into the pure reflectance spectra of the observed materials and the associated mixing coefficients. The formulation leads to a coupled matrix factorisation problem, with a number of useful constraints imposed by elementary physical properties of spectral mixing. In experiments with two benchmark datasets we show that the proposed approach delivers improved hyperspectral super-resolution.
Similar papers:
  • Hyperspectral Compressive Sensing Using Manifold-Structured Sparsity Prior [pdf] - Lei Zhang, Wei Wei, Yanning Zhang, Fei Li, Chunhua Shen, Qinfeng Shi
  • Pan-Sharpening With a Hyper-Laplacian Penalty [pdf] - Yiyong Jiang, Xinghao Ding, Delu Zeng, Yue Huang, John Paisley
  • Separating Fluorescent and Reflective Components by Using a Single Hyperspectral Image [pdf] - Yinqiang Zheng, Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf] - Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • RGB-Guided Hyperspectral Image Upsampling [pdf] - Hyeokhyen Kwon, Yu-Wing Tai
Exploring Causal Relationships in Visual Object Tracking [pdf]
Karel Lebeda, Simon Hadfield, Richard Bowden

Abstract: Causal relationships can often be found in visual object tracking between the motions of the camera and that of the tracked object. This object motion may be an effect of the camera motion, e.g. an unsteady handheld camera. But it may also be the cause, e.g. the cameraman framing the object. In this paper we explore these relationships, and provide statistical tools to detect and quantify them; these are based on transfer entropy and stem from information theory. The relationships are then exploited to make predictions about the object location. The approach is shown to be an excellent measure for describing such relationships. On the VOT2013 dataset the prediction accuracy is increased by 62 % over the best non-causal predictor. We show that the location predictions are robust to camera shake and sudden motion, which is invaluable for any tracking algorithm and demonstrate this by applying causal prediction to two state-of-the-art trackers. Both of them benefit, Struck gaining a 7 % accuracy and 22 % robustness increase on the VTB1.1 benchmark, becoming the new state-of-the-art.
Similar papers:
  • Understanding and Diagnosing Visual Tracking Systems [pdf] - Naiyan Wang, Jianping Shi, Dit-Yan Yeung, Jiaya Jia
  • Online Object Tracking With Proposal Selection [pdf] - Yang Hua, Karteek Alahari, Cordelia Schmid
  • Entropy-Based Latent Structured Output Prediction [pdf] - Diane Bouchacourt, Sebastian Nowozin, M. Pawan Kumar
  • TRIC-track: Tracking by Regression With Incrementally Learned Cascades [pdf] - Xiaomeng Wang, Michel Valstar, Brais Martinez, Muhammad Haris Khan, Tony Pridmore
  • A Versatile Learning-Based 3D Temporal Tracker: Scalable, Robust, Online [pdf] - David Joseph Tan, Federico Tombari, Slobodan Ilic, Nassir Navab
Fast and Accurate Head Pose Estimation via Random Projection Forests [pdf]
Donghoon Lee, Ming-Hsuan Yang, Songhwai Oh

Abstract: In this paper, we consider the problem of estimating the gaze direction of a person from a low-resolution image. Under this condition, reliably extracting facial features is very difficult. We propose a novel head pose estimation algorithm based on compressive sensing. Head image patches are mapped to a large feature space using the proposed extensive, yet efficient filter bank. The filter bank is designed to generate sparse responses of color and gradient information, which can be compressed using random projection, and classified by a random forest. Extensive experiments on challenging datasets show that the proposed algorithm performs favorably against the state-of-the-art methods on head pose estimation in low-resolution images degraded by noise, occlusion, and blurring.
Similar papers:
  • Rendering of Eyes for Eye-Shape Registration and Gaze Estimation [pdf] - Erroll Wood, Tadas Baltrusaitis, Xucong Zhang, Yusuke Sugano, Peter Robinson, Andreas Bulling
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
  • Person Recognition in Personal Photo Collections [pdf] - Seong Joon Oh, Rodrigo Benenson, Mario Fritz, Bernt Schiele
  • Uncovering Interactions and Interactors: Joint Estimation of Head, Body Orientation and F-Formations From Surveillance Videos [pdf] - Elisa Ricci, Jagannadan Varadarajan, Ramanathan Subramanian, Samuel Rota Bulo, Narendra Ahuja, Oswald Lanz
  • Robust Model-Based 3D Head Pose Estimation [pdf] - Gregory P. Meyer, Shalini Gupta, Iuri Frosio, Dikpal Reddy, Jan Kautz
Learning to Combine Mid-Level Cues for Object Proposal Generation [pdf]
Tom Lee, Sanja Fidler, Sven Dickinson

Abstract: In recent years, region proposals have replaced sliding windows in support of object recognition, offering more discriminating shape and appearance information through improved localization. One powerful approach for generating region proposals is based on minimizing parametric energy functions with parametric maxflow. In this paper, we introduce Parametric Min-Loss (PML), a novel structured learning framework for parametric energy functions. While PML is generally applicable to different domains, we use it in the context of region proposals to learn to combine a set of mid-level grouping cues to yield a small set of object region proposals with high recall. Our learning framework accounts for multiple diverse outputs, and is complemented by diversification seeds based on image location and color. This approach casts perceptual grouping and cue combination in a novel structured learning framework which yields baseline improvements on VOC 2012 and COCO 2014.
Similar papers:
  • Efficient Video Segmentation Using Parametric Graph Partitioning [pdf] - Chen-Ping Yu, Hieu Le, Gregory Zelinsky, Dimitris Samaras
  • Cascaded Sparse Spatial Bins for Efficient and Effective Generic Object Detection [pdf] - David Novotny, Jiri Matas
  • Oriented Object Proposals [pdf] - Shengfeng He, Rynson W.H. Lau
  • Fully Connected Object Proposals for Video Segmentation [pdf] - Federico Perazzi, Oliver Wang, Markus Gross, Alexander Sorkine-Hornung
  • The Middle Child Problem: Revisiting Parametric Min-Cut and Seeds for Object Proposals [pdf] - Ahmad Humayun, Fuxin Li, James M. Rehg
FollowMe: Efficient Online Min-Cost Flow Tracking With Bounded Memory and Computation [pdf]
Philip Lenz, Andreas Geiger, Raquel Urtasun

Abstract: One of the most popular approaches to multi-target tracking is tracking-by-detection. Current min-cost flow algorithms which solve the data association problem optimally have three main drawbacks: they are computationally expensive, they assume that the whole video is given as a batch, and they scale badly in memory and computation with the length of the video sequence. In this paper, we address each of these issues, resulting in a computationally and memory-bounded solution. First, we introduce a dynamic version of the successive shortest-path algorithm which solves the data association problem optimally while reusing computation, resulting in faster inference than standard solvers. Second, we address the optimal solution to the data association problem when dealing with an incoming stream of data (i.e., online setting). Finally, we present our main contribution which is an approximate online solution with bounded memory and computation which is capable of handling videos of arbitrary length while performing tracking in real time. We demonstrate the effectiveness of our algorithms on the KITTI and PETS2009 benchmarks and show state-of-the-art performance, while being significantly faster than existing solvers.
Similar papers:
  • Minimizing Human Effort in Interactive Tracking by Incremental Learning of Model Parameters [pdf] - Arridhana Ciptadi, James M. Rehg
  • Joint Probabilistic Data Association Revisited [pdf] - Seyed Hamid Rezatofighi, Anton Milan, Zhen Zhang, Qinfeng Shi, Anthony Dick, Ian Reid
  • Learning to Divide and Conquer for Online Multi-Target Tracking [pdf] - Francesco Solera, Simone Calderara, Rita Cucchiara
  • Learning to Track: Online Multi-Object Tracking by Decision Making [pdf] - Yu Xiang, Alexandre Alahi, Silvio Savarese
  • Near-Online Multi-Target Tracking With Aggregated Local Flow Descriptor [pdf] - Wongun Choi
Airborne Three-Dimensional Cloud Tomography [pdf]
Aviad Levis, Yoav Y. Schechner, Amit Aides, Anthony B. Davis

Abstract: orbit We seek to sense the three dimensional (3D) volumetric distribution of scatterers in a heterogenous medium. An important case study for such a medium is the atmosphere. Atmospheric contents and their role in Earths radiation balance have significant uncertainties with regards to scattering components: aerosols and clouds. Clouds, made of water droplets, also lead to local effects as precipitation and shadows. Our sensing approach is computational tomography using passive multi-angular imagery. For light-matter interaction that accounts for multiple-scattering, we use the 3D radiative transfer equation as a forward model. Volumetric recovery by inverting this model suffers from a computational bottleneck on large scales, which include many unknowns. Steps taken make this tomography tractable, without approximating the scattering order or angle range. IBC (x, ) I (x, ) J line t igh of s x0
Similar papers:
  • Depth Selective Camera: A Direct, On-Chip, Programmable Technique for Depth Selectivity in Photography [pdf] - Ryuichi Tadano, Adithya Kumar Pediredla, Ashok Veeraraghavan
  • Procedural Editing of 3D Building Point Clouds [pdf] - Ilke Demir, Daniel G. Aliaga, Bedrich Benes
  • On Linear Structure From Motion for Light Field Cameras [pdf] - Ole Johannsen, Antonin Sulc, Bastian Goldluecke
  • On Statistical Analysis of Neuroimages With Imperfect Registration [pdf] - Won Hwa Kim, Sathya N. Ravi, Sterling C. Johnson, Ozioma C. Okonkwo, Vikas Singh
  • Photometric Stereo in a Scattering Medium [pdf] - Zak Murez, Tali Treibitz, Ravi Ramamoorthi, David Kriegman
Live Repetition Counting [pdf]
Ofir Levy, Lior Wolf

Abstract: The task of counting the number of repetitions of approximately the same action in an input video sequence is addressed. The proposed method runs online and not on the complete pre-captured video. It analyzes sequentially blocks of 20 non-consecutive frames. The cycle length within each block is evaluated using a convolutional neural network and the information is then integrated over time. The entropy of the networks predictions is used in order to automatically start and stop the repetition counter and to select the appropriate time scale. Coupled with a region of interest detection mechanism, the method is robust enough to handle real world videos, even when the camera is moving. A unique property of our method is that it is shown to successfully train on entirely unrealistic data created by synthesizing moving random patches.
Similar papers:
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Human Action Recognition Using Factorized Spatio-Temporal Convolutional Networks [pdf] - Lin Sun, Kui Jia, Dit-Yan Yeung, Bertram E. Shi
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
3D Hand Pose Estimation Using Randomized Decision Forest With Segmentation Index Points [pdf]
Peiyi Li, Haibin Ling, Xi Li, Chunyuan Liao

Abstract: In this paper, we propose a real-time 3D hand pose estimation algorithm using the randomized decision forest framework. Our algorithm takes a depth image as input and generates a set of skeletal joints as output. Previous decision forest-based methods often give labels to all points in a point cloud at a very early stage and vote for the joint locations. By contrast, our algorithm only tracks a set of more flexible virtual landmark points, named segmentation index points (SIPs), before reaching the final decision at a leaf node. Roughly speaking, an SIP represents the centroid of a subset of skeletal joints, which are to be located at the leaves of the branch expanded from the SIP. Inspired by recent latent regression forest-based hand pose estimation framework (Tang et al. 2014), we integrate SIP into the framework with several important improvements: First, we devise a new forest growing strategy, whose decision is made using a randomized feature guided by SIPs. Second, we speed-up the training procedure since only SIPs, not the skeletal joints, are estimated at non-leaf nodes. Third, the experimental results on public benchmark datasets show clearly the advantage of the proposed algorithm over previous state-of-the-art methods, and our algorithm runs at 55.5 fps on a normal CPU without parallelism.
Similar papers:
  • Opening the Black Box: Hierarchical Sampling Optimization for Estimating Human Hand Pose [pdf] - Danhang Tang, Jonathan Taylor, Pushmeet Kohli, Cem Keskin, Tae-Kyun Kim, Jamie Shotton
  • Training a Feedback Loop for Hand Pose Estimation [pdf] - Markus Oberweger, Paul Wohlhart, Vincent Lepetit
  • Understanding Everyday Hands in Action From RGB-D Images [pdf] - Gregory Rogez, James S. Supancic III, Deva Ramanan
  • Depth-Based Hand Pose Estimation: Data, Methods, and Challenges [pdf] - James S. Supancic III, Gregory Rogez, Yi Yang, Jamie Shotton, Deva Ramanan
  • A Collaborative Filtering Approach to Real-Time Hand Pose Estimation [pdf] - Chiho Choi, Ayan Sinha, Joon Hee Choi, Sujin Jang, Karthik Ramani
A Data-Driven Metric for Comprehensive Evaluation of Saliency Models [pdf]
Jia Li, Changqun Xia, Yafei Song, Shu Fang, Xiaowu Chen

Abstract: In the past decades, hundreds of saliency models have been proposed for fixation prediction, along with dozens of evaluation metrics. However, existing metrics, which are often heuristically designed, may draw conflict conclusions in comparing saliency models. As a consequence, it becomes somehow confusing on the selection of metrics in comparing new models with state-of-the-arts. To address this problem, we propose a data-driven metric for comprehensive evaluation of saliency models. Instead of heuristically designing such a metric, we first conduct extensive subjective tests to find how saliency maps are assessed by the humanbeing. Based on the user data collected in the tests, nine representative evaluation metrics are directly compared by quantizing their performances in assessing saliency maps. Moreover, we propose to learn a data-driven metric by using Convolutional Neural Network. Compared with existing metrics, experimental results show that the data-driven metric performs the most consistently with the human-being in evaluating saliency maps as well as saliency models.
Similar papers:
  • Generic Promotion of Diffusion-Based Salient Object Detection [pdf] - Peng Jiang, Nuno Vasconcelos, Jingliang Peng
  • HARF: Hierarchy-Associated Rich Features for Salient Object Detection [pdf] - Wenbin Zou, Nikos Komodakis
  • Learning to Predict Saliency on Face Images [pdf] - Mai Xu, Yun Ren, Zulin Wang
  • Cluster-Based Point Set Saliency [pdf] - Flora Ponjou Tasse, Jiri Kosinka, Neil Dodgson
  • SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf] - Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao
Category-Blind Human Action Recognition: A Practical Recognition System [pdf]
Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu

Abstract: Existing human action recognition systems for 3D sequences obtained from the depth camera are designed to cope with only one action category, either single-person action or two-person interaction, and are difficult to be extended to scenarios where both action categories co-exist. In this paper, we propose the category-blind human recognition method (CHARM) which can recognize a human action without making assumptions of the action category. In our CHARM approach, we represent a human action (either a single-person action or a two-person interaction) class using a co-occurrence of motion primitives. Subsequently, we classify an action instance based on matching its motion primitive co-occurrence patterns to each class representation. The matching task is formulated as maximum clique problems. We conduct extensive evaluations of CHARM using three datasets for single-person actions, two-person interactions, and their mixtures. Experimental results show that CHARM performs favorably when compared with several state-of-the-art single-person action and two-person interaction based methods without making explicit assumptions of action category.
Similar papers:
  • Action Detection by Implicit Intentional Motion Clustering [pdf] - Wei Chen, Jason J. Corso
  • Action Localization in Videos Through Context Walk [pdf] - Khurram Soomro, Haroon Idrees, Mubarak Shah
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
Conformal and Low-Rank Sparse Representation for Image Restoration [pdf]
Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng

Abstract: of the learned dictionary becomes the key point for image restoration problem. Obtaining an appropriate dictionary is the key point when sparse representation is applied to computer vision or image processing problems such as image restoration. It is expected that preserving data structure during sparse coding and dictionary learning can enhance the recovery performance. However, many existing dictionary learning methods handle training samples individually, while missing relationships between samples, which result in dictionaries with redundant atoms but poor representation ability. In this paper, we propose a novel sparse representation approach called conformal and low-rank sparse representation (CLRSR) for image restoration problems. To achieve a more compact and representative dictionary, conformal property is introduced by preserving the angles of local geometry formed by neighboring samples in the feature space. Furthermore, imposing low-rank constraint on the coefficient matrix can lead more faithful subspaces and capture the global structure of data. We apply our CLRSR model to several image restoration tasks to demonstrate the effectiveness. Many approaches have been devoted to learning appropriate dictionaries. Some early dictionary learning methods, such as K-SVD algorithm [1], focus on reconstruction power of the dictionary, and have been applied to image restoration [9]. But these methods are dependent on a large training dataset. Besides, the dictionary size
Similar papers:
  • Person Re-Identification With Discriminatively Trained Viewpoint Invariant Dictionaries [pdf] - Srikrishna Karanam, Yang Li, Richard J. Radke
  • Variational Depth Superresolution Using Example-Based Edge Representations [pdf] - David Ferstl, Matthias Ruther, Horst Bischof
  • Convolutional Sparse Coding for Image Super-Resolution [pdf] - Shuhang Gu, Wangmeng Zuo, Qi Xie, Deyu Meng, Xiangchu Feng, Lei Zhang
  • Linearization to Nonlinear Learning for Visual Tracking [pdf] - Bo Ma, Hongwei Hu, Jianbing Shen, Yuping Zhang, Fatih Porikli
  • Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf] - Yuhui Quan, Yan Huang, Hui Ji
Dual-Feature Warping-Based Motion Model Estimation [pdf]
Shiwei Li, Lu Yuan, Jian Sun, Long Quan

Abstract: To break down the geometry assumptions of conventional motion models (e.g., homography, affine), warping-based motion model recently becomes popular and is adopted in many latest applications (e.g., image stitching, video stabilization). With high degrees of freedom, the accuracy of model heavily relies on data-terms (keypoint correspondences). In some low-texture environments (e.g., indoor) where keypoint feature is insufficient or unreliable, the warping model is often erroneously estimated. In this paper we propose a simple and effective approach by considering both keypoint and line segment correspondences as data-term. Line segment is a prominent feature in artificial environments and it can supply sufficient geometrical and structural information of scenes, which not only helps lead to a correct warp in low-texture condition, but also prevents the undesired distortion induced by warping. The combination aims to complement each other and benefit for a wider range of scenes. Our method is general and can be ported to many existing applications. Experiments demonstrate that using dual-feature yields more robust and accurate result especially for those low-texture images.
Similar papers:
  • Self-Occlusions and Disocclusions in Causal Video Object Segmentation [pdf] - Yanchao Yang, Ganesh Sundaramoorthi, Stefano Soatto
  • Robust RGB-D Odometry Using Point and Line Features [pdf] - Yan Lu, Dezhen Song
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
  • Synthesizing Illumination Mosaics From Internet Photo-Collections [pdf] - Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm
  • Integrating Dashcam Views Through Inter-Video Mapping [pdf] - Hsin-I Chen, Yi-Ling Chen, Wei-Tse Lee, Fan Wang, Bing-Yu Chen
Learning Parametric Distributions for Image Super-Resolution: Where Patch Matching Meets Sparse Coding [pdf]
Yongbo Li, Weisheng Dong, Guangming Shi, Xuemei Xie

Abstract: Existing approaches toward Image super-resolution (SR) is often either data-driven (e.g., based on internet-scale matching and web image retrieval) or model-based (e.g., formulated as an Maximizing a Posterior (MAP) estimation problem). The former is conceptually simple yet heuristic; while the latter is constrained by the fundamental limit of frequency aliasing. In this paper, we propose to develop a hybrid approach toward SR by combining those two lines of ideas. More specifically, the parameters underlying sparse distributions of desirable HR image patches are learned from a pair of LR image and retrieved HR images. Our hybrid approach can be interpreted as the first attempt of reconciling the difference between parametric and nonparametric models for low-level vision tasks. Experimental results show that the proposed hybrid SR method performs much better than existing state-of-the-art methods in terms of both subjective and objective image qualities.
Similar papers:
  • Variational Depth Superresolution Using Example-Based Edge Representations [pdf] - David Ferstl, Matthias Ruther, Horst Bischof
  • External Patch Prior Guided Internal Clustering for Image Denoising [pdf] - Fei Chen, Lei Zhang, Huimin Yu
  • Patch Group Based Nonlocal Self-Similarity Prior Learning for Image Denoising [pdf] - Jun Xu, Lei Zhang, Wangmeng Zuo, David Zhang, Xiangchu Feng
  • Convolutional Sparse Coding for Image Super-Resolution [pdf] - Shuhang Gu, Wangmeng Zuo, Qi Xie, Deyu Meng, Xiangchu Feng, Lei Zhang
  • Deep Networks for Image Super-Resolution With Sparse Prior [pdf] - Zhaowen Wang, Ding Liu, Jianchao Yang, Wei Han, Thomas Huang
Learning Semi-Supervised Representation Towards a Unified Optimization Framework for Semi-Supervised Learning [pdf]
Chun-Guang Li, Zhouchen Lin, Honggang Zhang, Jun Guo

Abstract: State of the art approaches for Semi-Supervised Learning (SSL) usually follow a two-stage framework constructing an affinity matrix from the data and then propagating the partial labels on this affinity matrix to infer those unknown labels. While such a two-stage framework has been successful in many applications, solving two subproblems separately only once is still suboptimal because it does not fully exploit the correlation between the affinity and the labels. In this paper, we formulate the two stages of SSL into a unified optimization framework, which learns both the affinity matrix and the unknown labels simultaneously. In the unified framework, both the given labels and the estimated labels are used to learn the affinity matrix and to infer the unknown labels. We solve the unified optimization problem via an alternating direction method of multipliers combined with label propagation. Extensive experiments on a synthetic data set and several benchmark data sets demonstrate the effectiveness of our approach.
Similar papers:
  • Semi-Supervised Zero-Shot Classification With Label Representation Learning [pdf] - Xin Li, Yuhong Guo, Dale Schuurmans
  • Probabilistic Label Relation Graphs With Ising Models [pdf] - Nan Ding, Jia Deng, Kevin P. Murphy, Hartmut Neven
  • A Matrix Decomposition Perspective to Multiple Graph Matching [pdf] - Junchi Yan, Hongteng Xu, Hongyuan Zha, Xiaokang Yang, Huanxi Liu, Stephen Chu
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • ML-MG: Multi-Label Learning With Missing Labels Using a Mixed Graph [pdf] - Baoyuan Wu, Siwei Lyu, Bernard Ghanem
Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf]
Sijin Li, Weichen Zhang, Antoni B. Chan

Abstract: This paper focuses on structured-output learning using deep neural networks for 3D human pose estimation from monocular images. Our network takes an image and 3D pose as inputs and outputs a score value, which is high when the image-pose pair matches and low otherwise. The network structure consists of a convolutional neural network for image feature extraction, followed by two sub-networks for transforming the image features and pose into a joint embedding. The score function is then the dot-product between the image and pose embeddings. The image-pose embedding and score function are jointly trained using a maximum-margin cost function. Our proposed framework can be interpreted as a special form of structured support vector machines where the joint feature space is discriminatively learned using deep neural networks. We test our framework on the Human3.6m dataset and obtain state-ofthe-art results compared to other recent methods. Finally, we present visualizations of the image-pose embedding space, demonstrating the network has learned a high-level embedding of body-orientation and pose-configuration.
Similar papers:
  • Learning Analysis-by-Synthesis for 6D Pose Estimation in RGB-D Images [pdf] - Alexander Krull, Eric Brachmann, Frank Michel, Michael Ying Yang, Stefan Gumhold, Carsten Rother
  • Training a Feedback Loop for Hand Pose Estimation [pdf] - Markus Oberweger, Paul Wohlhart, Vincent Lepetit
  • Flowing ConvNets for Human Pose Estimation in Videos [pdf] - Tomas Pfister, James Charles, Andrew Zisserman
  • Pose Induction for Novel Object Categories [pdf] - Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf] - Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang
Multi-Scale Learning for Low-Resolution Person Re-Identification [pdf]
Xiang Li, Wei-Shi Zheng, Xiaojuan Wang, Tao Xiang, Shaogang Gong

Abstract: In real world person re-identification (re-id), images of people captured at very different resolutions from different locations need be matched. Existing re-id models typically normalise all person images to the same size. However, a low-resolution (LR) image contains much less information about a person, and direct image scaling and simple size normalisation as done in conventional re-id methods cannot compensate for the loss of information. To solve this LR person re-id problem, we propose a novel joint multi-scale learning framework, termed joint multi-scale discriminant component analysis (JUDEA). The key component of this framework is a heterogeneous class mean discrepancy (HCMD) criterion for cross-scale image domain alignment, which is optimised simultaneously with discriminant modelling across multiple scales in the joint learning framework. Our experiments show that the proposed JUDEA framework outperforms existing representative reid methods as well as other related LR visual matching models applied for the LR person re-id problem.
Similar papers:
  • Person Re-Identification With Discriminatively Trained Viewpoint Invariant Dictionaries [pdf] - Srikrishna Karanam, Yang Li, Richard J. Radke
  • Efficient PSD Constrained Asymmetric Metric Learning for Person Re-Identification [pdf] - Shengcai Liao, Stan Z. Li
  • Scalable Person Re-Identification: A Benchmark [pdf] - Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian
  • Person Re-Identification Ranking Optimisation by Discriminant Context Information Analysis [pdf] - Jorge Garcia, Niki Martinel, Christian Micheloni, Alfredo Gardel
  • Partial Person Re-Identification [pdf] - Wei-Shi Zheng, Xiang Li, Tao Xiang, Shengcai Liao, Jianhuang Lai, Shaogang Gong
Nighttime Haze Removal With Glow and Multiple Light Colors [pdf]
Yu Li, Robby T. Tan, Michael S. Brown

Abstract: This paper focuses on dehazing nighttime images. Most existing dehazing methods use models that are formulated to describe haze in daytime. Daytime models assume a single uniform light color attributed to a light source not directly visible in the scene. Nighttime scenes, however, commonly include visible lights sources with varying colors. These light sources also often introduce noticeable amounts of glow that is not present in daytime haze. To address these effects, we introduce a new nighttime haze model that accounts for the varying light sources and their glow. Our model is a linear combination of three terms: the direct transmission, airlight and glow. The glow term represents light from the light sources that is scattered around before reaching the camera. Based on the model, we propose a framework that first reduces the effect of the glow in the image, resulting in a nighttime image that consists of direct transmission and airlight only. We then compute a spatially varying atmospheric light map that encodes light colors locally. This atmospheric map is used to predict the transmission, which we use to obtain our nighttime scene reflection image. We demonstrate the effectiveness of our nighttime haze model and correction method on a number of examples and compare our results with existing daytime and nighttime dehazing methods results.
Similar papers:
  • Depth Selective Camera: A Direct, On-Chip, Programmable Technique for Depth Selectivity in Photography [pdf] - Ryuichi Tadano, Adithya Kumar Pediredla, Ashok Veeraraghavan
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
  • Photometric Stereo in a Scattering Medium [pdf] - Zak Murez, Tali Treibitz, Ravi Ramamoorthi, David Kriegman
  • Photometric Stereo With Small Angular Variations [pdf] - Jian Wang, Yasuyuki Matsushita, Boxin Shi, Aswin C. Sankaranarayanan
  • On Linear Structure From Motion for Light Field Cameras [pdf] - Ole Johannsen, Antonin Sulc, Bastian Goldluecke
Selective Encoding for Recognizing Unreliably Localized Faces [pdf]
Ang Li, Vlad Morariu, Larry S. Davis

Abstract: Most existing face verification systems rely on precise face detection and registration. However, these two components are fallible under unconstrained scenarios (e.g., mobile face authentication) due to partial occlusions, pose variations, lighting conditions and limited view-angle coverage of mobile cameras. We address the unconstrained face verification problem by encoding face images directly without any explicit models of detection or registration. We propose a selective encoding framework which injects relevance information (e.g., foreground/background probabilities) into each cluster of a descriptor codebook. An additional selector component also discards distractive image patches and improves spatial robustness. We evaluate our framework using Gaussian mixture models and Fisher vectors on challenging face verification datasets. We apply selective encoding to Fisher vector features, which in our experiments degrade quickly with inaccurate face localization; our framework improves robustness with no extra test time computation. We also apply our approach to mobile based active face authentication task, demonstrating its utility in real scenarios.
Similar papers:
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • Deep Learning Face Attributes in the Wild [pdf] - Ziwei Liu, Ping Luo, Xiaogang Wang, Xiaoou Tang
  • Personalized Age Progression With Aging Dictionary [pdf] - Xiangbo Shu, Jinhui Tang, Hanjiang Lai, Luoqi Liu, Shuicheng Yan
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Simultaneous Local Binary Feature Learning and Encoding for Face Recognition [pdf] - Jiwen Lu, Venice Erin Liong, Jie Zhou
Semantic Video Entity Linking Based on Visual Content and Metadata [pdf]
Yuncheng Li, Xitong Yang, Jiebo Luo

Abstract: Video entity linking, which connects online videos to the related entities in a semantic knowledge base, can enable a wide variety of video based applications including video retrieval and video recommendation. Most existing systems for video entity linking rely on video metadata. In this paper, we propose to exploit video visual content to improve video entity linking. In the proposed framework, videos are first linked to entity candidates using a text-based method. Next, the entity candidates are verified and reranked according to visual content. In order to properly handle large variations in visual content matching, we propose to use Multiple Instance Metric Learning to learn a set to sequence metric for this specific matching problem. To evaluate the proposed framework, we collect and annotate 1912 videos crawled from the YouTube open API. Experiment results have shown consistent gains by the proposed framework over several strong baselines.
Similar papers:
  • Weakly-Supervised Alignment of Video With Text [pdf] - Piotr Bojanowski, Remi Lajugie, Edouard Grave, Francis Bach, Ivan Laptev, Jean Ponce, Cordelia Schmid
  • Unsupervised Extraction of Video Highlights Via Robust Recurrent Auto-Encoders [pdf] - Huan Yang, Baoyuan Wang, Stephen Lin, David Wipf, Minyi Guo, Baining Guo
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
Semi-Supervised Zero-Shot Classification With Label Representation Learning [pdf]
Xin Li, Yuhong Guo, Dale Schuurmans

Abstract: Given the challenge of gathering labeled training data, zero-shot classification, which transfers information from observed classes to recognize unseen classes, has become increasingly popular in the computer vision community. Most existing zero-shot learning methods require a user to first provide a set of semantic visual attributes for each class as side information before applying a two-step prediction procedure that introduces an intermediate attribute prediction problem. In this paper, we propose a novel zeroshot classification approach that automatically learns label embeddings from the input data in a semi-supervised large-margin learning framework. The proposed framework jointly considers multi-class classification over all classes (observed and unseen) and tackles the target prediction problem directly without introducing intermediate prediction problems. It also has the capacity to incorporate semantic label information from different sources when available. To evaluate the proposed approach, we conduct experiments on standard zero-shot data sets. The empirical results show the proposed approach outperforms existing state-of-the-art zero-shot learning methods.
Similar papers:
  • A Unified Multiplicative Framework for Attribute Learning [pdf] - Kongming Liang, Hong Chang, Shiguang Shan, Xilin Chen
  • Unsupervised Domain Adaptation for Zero-Shot Learning [pdf] - Elyor Kodirov, Tao Xiang, Zhenyong Fu, Shaogang Gong
  • ML-MG: Multi-Label Learning With Missing Labels Using a Mixed Graph [pdf] - Baoyuan Wu, Siwei Lyu, Bernard Ghanem
  • Predicting Deep Zero-Shot Convolutional Neural Networks Using Textual Descriptions [pdf] - Jimmy Lei Ba, Kevin Swersky, Sanja Fidler, Ruslan salakhutdinov
  • Zero-Shot Learning via Semantic Similarity Embedding [pdf] - Ziming Zhang, Venkatesh Saligrama
SPM-BP: Sped-up PatchMatch Belief Propagation for Continuous MRFs [pdf]
Yu Li, Dongbo Min, Michael S. Brown, Minh N. Do, Jiangbo Lu

Abstract: Markov random fields are widely used to model many computer vision problems that can be cast in an energy minimization framework composed of unary and pairwise potentials. While computationally tractable discrete optimizers such as Graph Cuts and belief propagation (BP) exist for multi-label discrete problems, they still face prohibitively high computational challenges when the labels reside in a huge or very densely sampled space. Integrating key ideas from PatchMatch of effective particle propagation and resampling, PatchMatch belief propagation (PMBP) has been demonstrated to have good performance in addressing continuous labeling problems and runs orders of magnitude faster than Particle BP (PBP). However, the quality of the PMBP solution is tightly coupled with the local window size, over which the raw data cost is aggregated to mitigate ambiguity in the data constraint. This dependency heavily influences the overall complexity, increasing linearly with the window size. This paper proposes a novel algorithm called sped-up PMBP (SPM-BP) to tackle this critical computational bottleneck and speeds up PMBP by 50-100 times. The crux of SPM-BP is on unifying efficient filter-based cost aggregation and message passing with PatchMatch-based particle generation in a highly effective way. Though simple in its formulation, SPM-BP achieves superior performance for sub-pixel accurate stereo and optical-flow on benchmark datasets when compared with more complex and taskspecific approa
Similar papers:
  • FlowNet: Learning Optical Flow With Convolutional Networks [pdf] - Alexey Dosovitskiy, Philipp Fischer, Eddy Ilg, Philip Hausser, Caner Hazirbas, Vladimir Golkov, Patrick van der Smagt, Daniel Cremers, Thomas Brox
  • A Deep Visual Correspondence Embedding Model for Stereo Matching Costs [pdf] - Zhuoyuan Chen, Xun Sun, Liang Wang, Yinan Yu, Chang Huang
  • Oriented Light-Field Windows for Scene Flow [pdf] - Pratul P. Srinivasan, Michael W. Tao, Ren Ng, Ravi Ramamoorthi
  • Massively Parallel Multiview Stereopsis by Surface Normal Diffusion [pdf] - Silvano Galliani, Katrin Lasinger, Konrad Schindler
  • Flow Fields: Dense Correspondence Fields for Highly Accurate Large Displacement Optical Flow Estimation [pdf] - Christian Bailer, Bertram Taetz, Didier Stricker
Temporal Subspace Clustering for Human Motion Segmentation [pdf]
Sheng Li, Kang Li, Yun Fu

Abstract: Output: Temporal segments Input: video sequence Temporal Regularization Subspace clustering is an effective technique for segmenting data drawn from multiple subspaces. However, for time series data (e.g., human motion), exploiting temporal information is still a challenging problem. We propose a novel temporal subspace clustering (TSC) approach in this paper. We improve the subspace clustering technique from two aspects. First, a temporal Laplacian regularization is designed, which encodes the sequential relationships in time series data. Second, to obtain expressive codings, we learn a non-negative dictionary from data. An efficient optimization algorithm is presented to jointly learn the representation codings and dictionary. After constructing an affinity graph using the codings, multiple temporal segments can be grouped via spectral clustering. Experimental results on three action and gesture datasets demonstrate the effectiveness of our approach. In particular, TSC significantly improves the clustering accuracy, compared to the state-of-the-art subspace clustering methods.
Similar papers:
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
  • Shape Interaction Matrix Revisited and Robustified: Efficient Subspace Clustering With Corrupted and Incomplete Data [pdf] - Pan Ji, Mathieu Salzmann, Hongdong Li
  • Low-Rank Tensor Constrained Multiview Subspace Clustering [pdf] - Changqing Zhang, Huazhu Fu, Si Liu, Guangcan Liu, Xiaochun Cao
  • Multi-View Subspace Clustering [pdf] - Hongchang Gao, Feiping Nie, Xuelong Li, Heng Huang
Two Birds, One Stone: Jointly Learning Binary Code for Large-Scale Face Image Retrieval and Attributes Prediction [pdf]
Yan Li, Ruiping Wang, Haomiao Liu, Huajie Jiang, Shiguang Shan, Xilin Chen

Abstract: We address the challenging large-scale content-based face image retrieval problem, intended as searching images based on the presence of specific subject, given one face image of him/her. To this end, one natural demand is a supervised binary code learning method. While the learned codes might be discriminating, people often have a further expectation that whether some semantic message (e.g., visual attributes) can be read from the human-incomprehensible codes. For this purpose, we propose a novel binary code learning framework by jointly encoding identity discriminability and a number of facial attributes into unified binary code. In this way, the learned binary codes can be applied to not only fine-grained face image retrieval, but also facial attributes prediction, which is the very innovation of this work, just like killing two birds with one stone. To evaluate the effectiveness of the proposed method, extensive experiments are conducted on a new purified large-scale web celebrity database, named CFW 60K, with abundant manual identity and attributes annotation, and experimental results exhibit the superiority of our method over state-of-theart.
Similar papers:
  • kNN Hashing With Factorized Neighborhood Representation [pdf] - Kun Ding, Chunlei Huo, Bin Fan, Chunhong Pan
  • An NMF Perspective on Binary Hashing [pdf] - Lopamudra Mukherjee, Sathya N. Ravi, Vamsi K. Ithapu, Tyler Holmes, Vikas Singh
  • Adaptive Hashing for Fast Similarity Search [pdf] - Fatih Cakir, Stan Sclaroff
  • Learning Binary Codes for Maximum Inner Product Search [pdf] - Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen
  • Alternating Co-Quantization for Cross-Modal Hashing [pdf] - Go Irie, Hiroyuki Arai, Yukinobu Taniguchi
A Unified Multiplicative Framework for Attribute Learning [pdf]
Kongming Liang, Hong Chang, Shiguang Shan, Xilin Chen

Abstract: Attributes are mid-level semantic properties of objects. Recent research has shown that visual attributes can benefit many traditional learning problems in computer vision community. However, attribute learning is still a challenging problem as the attributes may not always be predictable directly from input images and the variation of visual attributes is sometimes large across categories. In this paper, we propose a unified multiplicative framework for attribute learning, which tackles the key problems. Specifically, images and category information are jointly projected into a shared feature space, where the latent factors are disentangled and multiplied for attribute prediction. The resulting attribute classifier is category-specific instead of being shared by all categories. Moreover, our method can leverage auxiliary data to enhance the predictive ability of attribute classifiers, reducing the effort of instance-level attribute annotation to some extent. Experimental results show that our method achieves superior performance on both instance-level and category-level attribute prediction. For zero-shot learning based on attributes, our method significantly improves the state-of-theart performance on AwA dataset and achieves comparable performance on CUB dataset. lems (e.g., image search [15], object recognition [22] and face verification [16]). Moreover, they provide a proper way to address zero-shot classification [17] by transferring from seen classes to unseen classes
Similar papers:
  • Multi-Task Learning With Low Rank Attribute Embedding for Person Re-Identification [pdf] - Chi Su, Fan Yang, Shiliang Zhang, Qi Tian, Larry S. Davis, Wen Gao
  • Attributed Grammars for Joint Estimation of Human Attributes, Part and Pose [pdf] - Seyoung Park, Song-Chun Zhu
  • Just Noticeable Differences in Visual Attributes [pdf] - Aron Yu, Kristen Grauman
  • Discovering the Spatial Extent of Relative Attributes [pdf] - Fanyi Xiao, Yong Jae Lee
  • Learning Deep Representation With Large-Scale Attributes [pdf] - Wanli Ouyang, Hongyang Li, Xingyu Zeng, Xiaogang Wang
Human Parsing With Contextualized Convolutional Neural Network [pdf]
Xiaodan Liang, Chunyan Xu, Xiaohui Shen, Jianchao Yang, Si Liu, Jinhui Tang, Liang Lin, Shuicheng Yan

Abstract: In this work, we address the human parsing task with a novel Contextualized Convolutional Neural Network (CoCNN) architecture, which well integrates the cross-layer context, global image-level context, within-super-pixel context and cross-super-pixel neighborhood context into a unified network. Given an input human image, Co-CNN produces the pixel-wise categorization in an end-to-end way. First, the cross-layer context is captured by our basic localto-global-to-local structure, which hierarchically combines the global semantic information and the local fine details across different convolutional layers. Second, the global image-level label prediction is used as an auxiliary objective in the intermediate layer of the Co-CNN, and its outputs are further used for guiding the feature learning in subsequent convolutional layers to leverage the global imagelevel context. Finally, to further utilize the local super-pixel contexts, the within-super-pixel smoothing and cross-superpixel neighbourhood voting are formulated as natural subcomponents of the Co-CNN to achieve the local label consistency in both training and testing process. Comprehensive evaluations on two public datasets well demonstrate the significant superiority of our Co-CNN over other state-ofthe-arts for human parsing. In particular, the F-1 score on the large dataset [15] reaches 76.95% by Co-CNN, significantly higher than 62.81% and 64.38% by the state-of-theart algorithms, M-CNN [21] and ATR [15], respectively.
Similar papers:
  • ML-MG: Multi-Label Learning With Missing Labels Using a Mixed Graph [pdf] - Baoyuan Wu, Siwei Lyu, Bernard Ghanem
  • Conditional Random Fields as Recurrent Neural Networks [pdf] - Shuai Zheng, Sadeep Jayasumana, Bernardino Romera-Paredes, Vibhav Vineet, Zhizhong Su, Dalong Du, Chang Huang, Philip H. S. Torr
  • Learning Deconvolution Network for Semantic Segmentation [pdf] - Hyeonwoo Noh, Seunghoon Hong, Bohyung Han
  • Joint Object and Part Segmentation Using Deep Learned Potentials [pdf] - Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille
  • Semantic Image Segmentation via Deep Parsing Network [pdf] - Ziwei Liu, Xiaoxiao Li, Ping Luo, Chen-Change Loy, Xiaoou Tang
Towards Computational Baby Learning: A Weakly-Supervised Approach for Object Detection [pdf]
Xiaodan Liang, Si Liu, Yunchao Wei, Luoqi Liu, Liang Lin, Shuicheng Yan

Abstract: Intuitive observations show that a baby may inherently possess the capability of recognizing a new visual concept (e.g., chair, dog) by learning from only very few positive instances taught by parent(s) or others, and this recognition capability can be gradually further improved by exploring and/or interacting with the real instances in the physical world. Inspired by these observations, we propose a computational model for weakly-supervised object detection, based on prior knowledge modelling, exemplar learning and learning with video contexts. The prior knowledge is modeled with a pre-trained Convolutional Neural Network (CNN). When very few instances of a new concept are given, an initial concept detector is built by exemplar learning over the deep features the pre-trained CNN. The well-designed tracking solution is then used to discover more diverse instances from the massive online weakly labeled videos. Once a positive instance is detected/identified with high score in each video, more instances possibly from different view-angles and/or different distances are tracked and accumulated. Then the concept detector can be finetuned based on these new instances. This process can be repeated again and again till we obtain a very mature concept detector. Extensive experiments on Pascal VOC-07/10/12 object detection datasets [9] well demonstrate the effectiveness of our framework. It can beat the state-of-the-art full-training based performances by learning from very few sample
Similar papers:
  • Automatic Concept Discovery From Parallel Text and Visual Corpora [pdf] - Chen Sun, Chuang Gan, Ram Nevatia
  • Relaxed Multiple-Instance SVM With Application to Object Discovery [pdf] - Xinggang Wang, Zhuotun Zhu, Cong Yao, Xiang Bai
  • Selecting Relevant Web Trained Concepts for Automated Event Retrieval [pdf] - Bharat Singh, Xintong Han, Zhe Wu, Vlad I. Morariu, Larry S. Davis
  • A Self-Paced Multiple-Instance Learning Framework for Co-Saliency Detection [pdf] - Dingwen Zhang, Deyu Meng, Chao Li, Lu Jiang, Qian Zhao, Junwei Han
  • Simpler Non-Parametric Methods Provide as Good or Better Results to Multiple-Instance Learning [pdf] - Ragav Venkatesan, Parag Chandakkar, Baoxin Li
Efficient PSD Constrained Asymmetric Metric Learning for Person Re-Identification [pdf]
Shengcai Liao, Stan Z. Li

Abstract: approaches have been proposed for person re-identification [4], which greatly advance this field. Person re-identification is becoming a hot research topic due to its value in both machine learning research and video surveillance applications. For this challenging problem, distance metric learning is shown to be effective in matching person images. However, existing approaches either require a heavy computation due to the positive semidefinite (PSD) constraint, or ignore the PSD constraint and learn a free distance function that makes the learned metric potentially noisy. We argue that the PSD constraint provides a useful regularization to smooth the solution of the metric, and hence the learned metric is more robust than without the PSD constraint. Another problem with metric learning algorithms is that the number of positive sample pairs is very limited, and the learning process is largely dominated by the large amount of negative sample pairs. To address the above issues, we derive a logistic metric learning approach with the PSD constraint and an asymmetric sample weighting strategy. Besides, we successfully apply the accelerated proximal gradient approach to find a global minimum solution of the proposed formulation, with a convergence rate of O(1/t2 ) where t is the number of iterations. The proposed algorithm termed MLAPG is shown to be computationally efficient and able to perform low rank selection. We applied the proposed method for person re-identification, achie
Similar papers:
  • A Supervised Low-Rank Method for Learning Invariant Subspaces [pdf] - Farzad Siyahjani, Ranya Almohsen, Sinan Sabri, Gianfranco Doretto
  • Multi-Task Learning With Low Rank Attribute Embedding for Person Re-Identification [pdf] - Chi Su, Fan Yang, Shiliang Zhang, Qi Tian, Larry S. Davis, Wen Gao
  • Multi-Scale Learning for Low-Resolution Person Re-Identification [pdf] - Xiang Li, Wei-Shi Zheng, Xiaojuan Wang, Tao Xiang, Shaogang Gong
  • Scalable Nonlinear Embeddings for Semantic Category-Based Image Retrieval [pdf] - Gaurav Sharma, Bernt Schiele
  • Scalable Person Re-Identification: A Benchmark [pdf] - Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian
Video Super-Resolution via Deep Draft-Ensemble Learning [pdf]
Renjie Liao, Xin Tao, Ruiyu Li, Ziyang Ma, Jiaya Jia

Abstract: We propose a new direction for fast video superresolution (VideoSR) via a SR draft ensemble, which is defined as the set of high-resolution patch candidates before final image deconvolution. Our method contains two main components i.e., SR draft ensemble generation and its optimal reconstruction. The first component is to renovate traditional feedforward reconstruction pipeline and greatly enhance its ability to compute different super-resolution results considering large motion variation and possible errors arising in this process. Then we combine SR drafts through the nonlinear process in a deep convolutional neural network (CNN). We analyze why this framework is proposed and explain its unique advantages compared to previous iterative methods to update different modules in passes. Promising experimental results are shown on natural video sequences.
Similar papers:
  • Recurrent Network Models for Human Dynamics [pdf] - Katerina Fragkiadaki, Sergey Levine, Panna Felsen, Jitendra Malik
  • Self-Occlusions and Disocclusions in Causal Video Object Segmentation [pdf] - Yanchao Yang, Ganesh Sundaramoorthi, Stefano Soatto
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
  • Face Flow [pdf] - Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou
  • Dense Optical Flow Prediction From a Static Image [pdf] - Jacob Walker, Abhinav Gupta, Martial Hebert
Bilinear CNN Models for Fine-Grained Visual Recognition [pdf]
Tsung-Yu Lin, Aruni RoyChowdhury, Subhransu Maji

Abstract: We propose bilinear models, a recognition architecture that consists of two feature extractors whose outputs are multiplied using outer product at each location of the image and pooled to obtain an image descriptor. This architecture can model local pairwise feature interactions in a translationally invariant manner which is particularly useful for fine-grained categorization. It also generalizes various orderless texture descriptors such as the Fisher vector, VLAD and O2P. We present experiments with bilinear models where the feature extractors are based on convolutional neural networks. The bilinear form simplifies gradient computation and allows end-to-end training of both networks using image labels only. Using networks initialized from the ImageNet dataset followed by domain specific fine-tuning we obtain 84.1% accuracy of the CUB200-2011 dataset requiring only category labels at training time. We present experiments and visualizations that analyze the effects of fine-tuning and the choice two networks on the speed and accuracy of the models. Results show that the architecture compares favorably to the existing state of the art on a number of fine-grained datasets while being substantially simpler and easier to train. Moreover, our most accurate model is fairly efficient running at 8 frames/sec on a NVIDIA Tesla K40 GPU. The source code for the complete system will be made available at http://vis-www.cs.umass.edu/bcnn
Similar papers:
  • Projection Bank: From High-Dimensional Data to Medium-Length Binary Codes [pdf] - Li Liu, Mengyang Yu, Ling Shao
  • Task-Driven Feature Pooling for Image Classification [pdf] - Guo-Sen Xie, Xu-Yao Zhang, Xiangbo Shu, Shuicheng Yan, Cheng-Lin Liu
  • Aggregating Local Deep Features for Image Retrieval [pdf] - Artem Babenko, Victor Lempitsky
  • Fast R-CNN [pdf] - Ross Girshick
  • Multi-Scale Recognition With DAG-CNNs [pdf] - Songfan Yang, Deva Ramanan
Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf]
Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang

Abstract: Wearable cameras, such as Google Glass and Go Pro, enable video data collection over larger areas and from different views. In this paper, we tackle a new problem of locating the co-interest person (CIP), i.e., the one who draws attention from most camera wearers, from temporally synchronized videos taken by multiple wearable cameras. Our basic idea is to exploit the motion patterns of people and use them to correlate the persons across different videos, instead of performing appearance-based matching as in traditional video co-segmentation/localization. This way, we can identify CIP even if a group of people with similar appearance are present in the view. More specifically, we detect a set of persons on each frame as the candidates of the CIP and then build a Conditional Random Field (CRF) model to select the one with consistent motion patterns in different videos and high spacial-temporal consistency in each video. We collect three sets of wearable-camera videos for testing the proposed algorithm. All the involved people have similar appearances in the collected videos and the experiments demonstrate the effectiveness of the proposed algorithm.
Similar papers:
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
  • Semantic Video Entity Linking Based on Visual Content and Metadata [pdf] - Yuncheng Li, Xitong Yang, Jiebo Luo
  • Video Restoration Against Yin-Yang Phasing [pdf] - Xiaolin Wu, Zhenhao Li, Xiaowei Deng
  • Integrating Dashcam Views Through Inter-Video Mapping [pdf] - Hsin-I Chen, Yi-Ling Chen, Wei-Tse Lee, Fan Wang, Bing-Yu Chen
Depth Recovery From Light Field Using Focal Stack Symmetry [pdf]
Haiting Lin, Can Chen, Sing Bing Kang, Jingyi Yu

Abstract: We describe a technique to recover depth from a light field (LF) using two proposed features of the LF focal stack. One feature is the property that non-occluding pixels exhibit symmetry along the focal depth dimension centered at the in-focus slice. The other is a data consistency measure based on analysis-by-synthesis, i.e., the difference between the synthesized focal stack given the hypothesized depth map and that from the LF. These terms are used in an iterative optimization framework to extract scene depth. Experimental results on real Lytro and Raytrix data demonstrate that our technique outperforms state-of-the-art solutions and is significantly more robust to noise and undersampling. yu@eecis.udel.edu Scene Original BM3D Ours on original GC on original GC on denoised Figure 1. Noise handling (Lytro raw data extracted by [2]). Traditional stereo matching approaches either use a large smoothness term or first denoise the input [8, 41]. Both have the effect of blurring boundaries. Our technique is able to recover fine details without oversmoothing using the original noisy input.
Similar papers:
  • SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf] - Kensei Jo, Mohit Gupta, Shree K. Nayar
  • Depth Map Estimation and Colorization of Anaglyph Images Using Local Color Prior and Reverse Intensity Distribution [pdf] - Williem, Ramesh Raskar, In Kyu Park
  • MAP Disparity Estimation Using Hidden Markov Trees [pdf] - Eric T. Psota, Jedrzej Kowalczuk, Mateusz Mittek, Lance C. Perez
  • Blur-Aware Disparity Estimation From Defocus Stereo Images [pdf] - Ching-Hui Chen, Hui Zhou, Timo Ahonen
  • Extended Depth of Field Catadioptric Imaging Using Focal Sweep [pdf] - Ryunosuke Yokoya, Shree K. Nayar
A Spatio-Temporal Appearance Representation for Viceo-Based Pedestrian Re-Identification [pdf]
Kan Liu, Bingpeng Ma, Wei Zhang, Rui Huang

Abstract: Pedestrian re-identification is a difficult problem due to the large variations in a persons appearance caused by different poses and viewpoints, illumination changes, and occlusions. Spatial alignment is commonly used to address these issues by treating the appearance of different body parts independently. However, a body part can also appear differently during different phases of an action. In this paper we consider the temporal alignment problem, in addition to the spatial one, and propose a new approach that takes the video of a walking person as input and builds a spatio-temporal appearance representation for pedestrian re-identification. Particularly, given a video sequence we exploit the periodicity exhibited by a walking person to generate a spatio-temporal body-action model, which consists of a series of body-action units corresponding to certain action primitives of certain body parts. Fisher vectors are learned and extracted from individual body-action units and concatenated into the final representation of the walking person. Unlike previous spatio-temporal features that only take into account local dynamic appearance information, our representation aligns the spatio-temporal appearance of a pedestrian globally. Extensive experiments on public datasets show the effectiveness of our approach compared with the state of the art.
Similar papers:
  • Person Re-Identification With Discriminatively Trained Viewpoint Invariant Dictionaries [pdf] - Srikrishna Karanam, Yang Li, Richard J. Radke
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
  • Scalable Person Re-Identification: A Benchmark [pdf] - Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian
  • Recurrent Network Models for Human Dynamics [pdf] - Katerina Fragkiadaki, Sergey Levine, Panna Felsen, Jitendra Malik
  • Human Pose Estimation in Videos [pdf] - Dong Zhang, Mubarak Shah
Bayesian Model Adaptation for Crowd Counts [pdf]
Bo Liu, Nuno Vasconcelos

Abstract: The problem of transfer learning is considered in the domain of crowd counting. A solution based on Bayesian model adaptation of Gaussian processes is proposed. This is shown to produce intuitive model updates, which are tractable, and lead to an adapted model (predictive distribution) that accounts for all information in both training and adaptation data. The new adaptation procedure achieves significant gains over previous approaches, based on multi-task learning, while requiring much less computation to deploy. This makes it particularly suited for the problem of expanding the capacity of crowd counting camera networks. A large video dataset for the evaluation of adaptation approaches to crowd counting is also introduced. This contains a number of adaptation tasks, involving information transfer across video collected by 1) a single camera under different scene conditions (different times of the day) and 2) video collected from different cameras. Evaluation of the proposed model adaptation procedure in this dataset shows good performance in realistic operating conditions.
Similar papers:
  • Simultaneous Deep Transfer Across Domains and Tasks [pdf] - Eric Tzeng, Judy Hoffman, Trevor Darrell, Kate Saenko
  • COUNT Forest: CO-Voting Uncertain Number of Targets Using Random Forest for Crowd Density Estimation [pdf] - Viet-Quoc Pham, Tatsuo Kozakaya, Osamu Yamaguchi, Ryuzo Okada
  • Unsupervised Domain Adaptation for Zero-Shot Learning [pdf] - Elyor Kodirov, Tao Xiang, Zhenyong Fu, Shaogang Gong
  • Unsupervised Domain Adaptation With Imbalanced Cross-Domain Data [pdf] - Tzu Ming Harry Hsu, Wei Yu Chen, Cheng-An Hou, Yao-Hung Hubert Tsai, Yi-Ren Yeh, Yu-Chiang Frank Wang
  • Per-Sample Kernel Adaptation for Visual Recognition and Grouping [pdf] - Borislav Antic, Bjorn Ommer
Box Aggregation for Proposal Decimation: Last Mile of Object Detection [pdf]
Shu Liu, Cewu Lu, Jiaya Jia

Abstract: Regions-with-convolutional-neural-network (RCNN) is now a commonly employed object detection pipeline. Its main steps, i.e., proposal generation and convolutional neural network (CNN) feature extraction, have been intensively investigated. We focus on the last step of the system to aggregate thousands of scored box proposals into final object prediction, which we call proposal decimation. We show this step can be enhanced with a very simple box aggregation function by considering statistical properties of proposals with respect to ground truth objects. Our method is with extremely light-weight computation, while it yields an improvement of 3.7% in mAP on PASCAL VOC 2007 test. We explain why it works using some statistics in this paper.
Similar papers:
  • Boosting Object Proposals: From Pascal to COCO [pdf] - Jordi Pont-Tuset, Luc Van Gool
  • DeepProposal: Hunting Objects by Cascading Deep Convolutional Layers [pdf] - Amir Ghodrati, Ali Diba, Marco Pedersoli, Tinne Tuytelaars, Luc Van Gool
  • Fully Connected Object Proposals for Video Segmentation [pdf] - Federico Perazzi, Oliver Wang, Markus Gross, Alexander Sorkine-Hornung
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • Cascaded Sparse Spatial Bins for Efficient and Effective Generic Object Detection [pdf] - David Novotny, Jiri Matas
Deep Learning Face Attributes in the Wild [pdf]
Ziwei Liu, Ping Luo, Xiaogang Wang, Xiaoou Tang

Abstract: Predicting face attributes in the wild is challenging due to complex face variations. We propose a novel deep learning framework for attribute prediction in the wild. It cascades two CNNs, LNet and ANet, which are finetuned jointly with attribute tags, but pre-trained differently. LNet is pre-trained by massive general object categories for face localization, while ANet is pre-trained by massive face identities for attribute prediction. This framework not only outperforms the state-of-the-art with a large margin, but also reveals valuable facts on learning face representation. (1) It shows how the performances of face localization (LNet) and attribute prediction (ANet) can be improved by different pre-training strategies. (2) It reveals that although the filters of LNet are fine-tuned only with imagelevel attribute tags, their response maps over entire images have strong indication of face locations. This fact enables training LNet for face localization with only image-level annotations, but without face bounding boxes or landmarks, which are required by all attribute recognition works. (3) It also demonstrates that the high-level hidden neurons of ANet automatically discover semantic concepts after pretraining with massive face identities, and such concepts are significantly enriched after fine-tuning with attribute tags. Each attribute can be well explained with a sparse linear combination of these concepts.
Similar papers:
  • Learning Deep Representation With Large-Scale Attributes [pdf] - Wanli Ouyang, Hongyang Li, Xingyu Zeng, Xiaogang Wang
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • Simultaneous Local Binary Feature Learning and Encoding for Face Recognition [pdf] - Jiwen Lu, Venice Erin Liong, Jie Zhou
  • A Unified Multiplicative Framework for Attribute Learning [pdf] - Kongming Liang, Hong Chang, Shiguang Shan, Xilin Chen
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
Higher-Order CRF Structural Segmentation of 3D Reconstructed Surfaces [pdf]
Jingbo Liu, Jinglu Wang, Tian Fang, Chiew-Lan Tai, Long Quan

Abstract: In this paper, we propose a structural segmentation algorithm to partition multi-view stereo reconstructed surfaces of large-scale urban environments into structural segments. Each segment corresponds to a structural component describable by a surface primitive of up to the second order. This segmentation is for use in subsequent urban object modeling, vectorization, and recognition. To overcome the high geometrical and topological noise levels in the 3D reconstructed urban surfaces, we formulate the structural segmentation as a higher-order Conditional Random Field (CRF) labeling problem. It not only incorporates classical lower-order 2D and 3D local cues, but also encodes contextual geometric regularities to disambiguate the noisy local cues. A general higher-order CRF is difficult to solve. We develop a bottom-up progressive approach through a patch-based surface representation, which iteratively evolves from the initial mesh triangles to the final segmentation. Each iteration alternates between performing a prior discovery step, which finds the contextual regularities of the patch-based representation, and an inference step that leverages the regularities as higher-order priors to construct a more stable and regular segmentation. The efficiency and robustness of the proposed method is extensively demonstrated on real reconstruction models, yielding significantly better performance than classical mesh segmentation methods.
Similar papers:
  • Procedural Editing of 3D Building Point Clouds [pdf] - Ilke Demir, Daniel G. Aliaga, Bedrich Benes
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • Peeking Template Matching for Depth Extension [pdf] - Simon Korman, Eyal Ofek, Shai Avidan
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Joint Camera Clustering and Surface Segmentation for Large-Scale Multi-View Stereo [pdf] - Runze Zhang, Shiwei Li, Tian Fang, Siyu Zhu, Long Quan
Illumination Robust Color Naming via Label Propagation [pdf]
Yuanliu liu, Zejian Yuan, Badong Chen, Jianru Xue, Nanning Zheng

Abstract: Color composition is an important property for many computer vision tasks like image retrieval and object classification. In this paper we address the problem of inferring the color composition of the intrinsic reflectance of objects, where the shadows and highlights may change the observed color dramatically. We achieve this through color label propagation without recovering the intrinsic reflectance beforehand. Specifically, the color labels are propagated between regions sharing the same reflectance, and the direction of propagation is promoted to be from regions under full illumination and normal view angles to abnormal regions. We detect shadowed and highlighted regions as well as pairs of regions that have similar reflectance. A joint inference process is adopted to trim the inconsistent identities and connections. For evaluation we collect three datasets of images under noticeable highlights and shadows. Experimental results show that our model can effectively describe the color composition of real-world images.
Similar papers:
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
  • Convolutional Color Constancy [pdf] - Jonathan T. Barron
  • Beyond White: Ground Truth Colors for Color Constancy Correction [pdf] - Dongliang Cheng, Brian Price, Scott Cohen, Michael S. Brown
  • Leave-One-Out Kernel Optimization for Shadow Detection [pdf] - Tomas F. Yago Vicente, Minh Hoai, Dimitris Samaras
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
Multi-View Complementary Hash Tables for Nearest Neighbor Search [pdf]
Xianglong Liu, Lei Huang, Cheng Deng, Jiwen Lu, Bo Lang

Abstract: Recent years have witnessed the success of hashing techniques in fast nearest neighbor search. In practice many applications (e.g., visual search, object detection, image matching, etc.) have enjoyed the benefits of complementary hash tables and information fusion over multiple views. However, most of prior research mainly focused on compact hash code cleaning, and rare work studies how to build multiple complementary hash tables, much less to adaptively integrate information stemming from multiple views. In this paper we first present a novel multi-view complementary hash table method that learns complementary hash tables from the data with multiple views. For single multiview table, using exemplar based feature fusion, we approximate the inherent data similarities with a low-rank matrix, and learn discriminative hash functions in an efficient way. To build complementary tables and meanwhile maintain scalable training and fast out-of-sample extension, an exemplar reweighting scheme is introduced to update the induced low-rank similarity in the sequential table construction framework, which indeed brings mutual benefits between tables by placing greater importance on exemplars shared by mis-separated neighbors. Extensive experiments on three large-scale image datasets demonstrate that the proposed method significantly outperforms various naive solutions and state-of-the-art multi-table methods.
Similar papers:
  • Two Birds, One Stone: Jointly Learning Binary Code for Large-Scale Face Image Retrieval and Attributes Prediction [pdf] - Yan Li, Ruiping Wang, Haomiao Liu, Huajie Jiang, Shiguang Shan, Xilin Chen
  • Alternating Co-Quantization for Cross-Modal Hashing [pdf] - Go Irie, Hiroyuki Arai, Yukinobu Taniguchi
  • Learning Binary Codes for Maximum Inner Product Search [pdf] - Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen
  • kNN Hashing With Factorized Neighborhood Representation [pdf] - Kun Ding, Chunlei Huo, Bin Fan, Chunhong Pan
  • Adaptive Hashing for Fast Similarity Search [pdf] - Fatih Cakir, Stan Sclaroff
PatchMatch-Based Automatic Lattice Detection for Near-Regular Textures [pdf]
Siying Liu, Tian-Tsong Ng, Kalyan Sunkavalli, Minh N. Do, Eli Shechtman, Nathan Carr

Abstract: In this work, we investigate the problem of automatically inferring the lattice structure of near-regular textures (NRT) in real-world images. Our technique leverages the PatchMatch algorithm for finding k-nearest-neighbor (kNN) correspondences in an image. We use these kNNs to recover an initial estimate of the 2D wallpaper basis vectors, and seed vertices of the texture lattice. We iteratively expand this lattice by solving an MRF optimization problem. We show that we can discretize the space of good solutions for the MRF using the kNNs, allowing us to efficiently and accurately optimize the MRF energy function using the Particle Belief Propagation algorithm. We demonstrate our technique on a benchmark NRT dataset containing a wide range of images with geometric and photometric variations, and show that our method clearly outperforms the state of the art in terms of both texel detection rate and texel localization score.
Similar papers:
  • Variational PatchMatch MultiView Reconstruction and Refinement [pdf] - Philipp Heise, Brian Jensen, Sebastian Klose, Alois Knoll
  • Massively Parallel Multiview Stereopsis by Surface Normal Diffusion [pdf] - Silvano Galliani, Katrin Lasinger, Konrad Schindler
  • Mode-Seeking on Hypergraphs for Robust Geometric Model Fitting [pdf] - Hanzi Wang, Guobao Xiao, Yan Yan, David Suter
  • SPM-BP: Sped-up PatchMatch Belief Propagation for Continuous MRFs [pdf] - Yu Li, Dongbo Min, Michael S. Brown, Minh N. Do, Jiangbo Lu
  • Example-Based Modeling of Facial Texture From Deficient Data [pdf] - Arnaud Dessein, William A. P. Smith, Richard C. Wilson, Edwin R. Hancock
Projection Bank: From High-Dimensional Data to Medium-Length Binary Codes [pdf]
Li Liu, Mengyang Yu, Ling Shao

Abstract: Recently, very high-dimensional feature representations, e.g., Fisher Vector, have achieved excellent performance for visual recognition and retrieval. However, these lengthy representations always cause extremely heavy computational and storage costs and even become unfeasible in some large-scale applications. A few existing techniques can transfer very high-dimensional data into binary codes, but they still require the reduced code length to be relatively long to maintain acceptable accuracies. To target a better balance between computational efficiency and accuracies, in this paper, we propose a novel embedding method called Binary Projection Bank (BPB), which can effectively reduce the very high-dimensional representations to medium-dimensional binary codes without sacrificing accuracies. Instead of using conventional single linear or bilinear projections, the proposed method learns a bank of small projections via the max-margin constraint to optimally preserve the intrinsic data similarity. We have systematically evaluated the proposed method on three datasets: Flickr 1M, ILSVR2010 and UCF101, showing competitive retrieval and recognition accuracies compared with stateof-the-art approaches, but with a significantly smaller memory footprint and lower coding complexity.
Similar papers:
  • Two Birds, One Stone: Jointly Learning Binary Code for Large-Scale Face Image Retrieval and Attributes Prediction [pdf] - Yan Li, Ruiping Wang, Haomiao Liu, Huajie Jiang, Shiguang Shan, Xilin Chen
  • An NMF Perspective on Binary Hashing [pdf] - Lopamudra Mukherjee, Sathya N. Ravi, Vamsi K. Ithapu, Tyler Holmes, Vikas Singh
  • Top Rank Supervised Binary Coding for Visual Search [pdf] - Dongjin Song, Wei Liu, Rongrong Ji, David A. Meyer, John R. Smith
  • Alternating Co-Quantization for Cross-Modal Hashing [pdf] - Go Irie, Hiroyuki Arai, Yukinobu Taniguchi
  • Learning Binary Codes for Maximum Inner Product Search [pdf] - Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen
Semantic Image Segmentation via Deep Parsing Network [pdf]
Ziwei Liu, Xiaoxiao Li, Ping Luo, Chen-Change Loy, Xiaoou Tang

Abstract: This paper addresses semantic image segmentation by incorporating rich information into Markov Random Field (MRF), including high-order relations and mixture of label contexts. Unlike previous works that optimized MRFs using iterative algorithm, we solve MRF by proposing a Convolutional Neural Network (CNN), namely Deep Parsing Network (DPN)1 , which enables deterministic end-toend computation in a single forward pass. Specifically, DPN extends a contemporary CNN architecture to model unary terms and additional layers are carefully devised to approximate the mean field algorithm (MF) for pairwise terms. It has several appealing properties. First, different from the recent works that combined CNN and MRF, where many iterations of MF were required for each training image during back-propagation, DPN is able to achieve high performance by approximating one iteration of MF. Second, DPN represents various types of pairwise terms, making many existing works as its special cases. Third, DPN makes MF easier to be parallelized and speeded up in Graphical Processing Unit (GPU). DPN is thoroughly evaluated on the PASCAL VOC 2012 dataset, where a single DPN model yields a new state-of-the-art segmentation accuracy of 77.5%.
Similar papers:
  • Joint Object and Part Segmentation Using Deep Learned Potentials [pdf] - Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille
  • Human Parsing With Contextualized Convolutional Neural Network [pdf] - Xiaodan Liang, Chunyan Xu, Xiaohui Shen, Jianchao Yang, Si Liu, Jinhui Tang, Liang Lin, Shuicheng Yan
  • Semantic Segmentation With Object Clique Potential [pdf] - Xiaojuan Qi, Jianping Shi, Shu Liu, Renjie Liao, Jiaya Jia
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
  • Conditional Random Fields as Recurrent Neural Networks [pdf] - Shuai Zheng, Sadeep Jayasumana, Bernardino Romera-Paredes, Vibhav Vineet, Zhizhong Su, Dalong Du, Chang Huang, Philip H. S. Torr
Joint Optimization of Segmentation and Color Clustering [pdf]
Ekaterina Lobacheva, Olga Veksler, Yuri Boykov

Abstract: Binary energy optimization is a popular approach for segmenting an image into foreground/background regions. To model region appearance, color, a relatively high dimensional feature, should be handled effectively. A full color histogram is usually too sparse to be reliable. One approach is to reduce dimensionality by color space clustering. Another popular approach is to fit GMMs for soft color space clustering. These approaches work well when the foreground/background are sufficiently distinct. In cases of more subtle difference in appearance, both approaches may reduce or even eliminate foreground/background distinction. This happens because either color clustering is performed completely independently from segmentation, as a preprocessing step (in clustering), or independently for the foreground and independently for the background (in GMM). We propose to make clustering an integral part of segmentation, by including a new clustering term in the energy. Our energy favors clusterings that make foreground/background appearance more distinct. Exact optimization is not feasible, therefore we develop an approximate algorithm. We show the advantage of including the color clustering term into the energy function on camouflage images, as well as standard segmentation datasets.
Similar papers:
  • Volumetric Bias in Segmentation and Reconstruction: Secrets and Solutions [pdf] - Yuri Boykov, Hossam Isack, Carl Olsson, Ismail Ben Ayed
  • Temporal Subspace Clustering for Human Motion Segmentation [pdf] - Sheng Li, Kang Li, Yun Fu
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
  • Secrets of GrabCut and Kernel K-Means [pdf] - Meng Tang, Ismail Ben Ayed, Dmitrii Marin, Yuri Boykov
Multi-Class Multi-Annotator Active Learning With Robust Gaussian Process for Visual Recognition [pdf]
Chengjiang Long, Gang Hua

Abstract: Active learning is an effective way to relieve the tedious work of manual annotation in many applications of visual recognition. However, less research attention has been focused on multi-class active learning. In this paper, we propose a novel Gaussian process classifier model with multiple annotators for multi-class visual recognition. Expectation propagation (EP) is adopted for efficient approximate Bayesian inference of our probabilistic model for classification. Based on the EP approximation inference, a generalized Expectation Maximization (GEM) algorithm is derived to estimate both the parameters for instances and the quality of each individual annotator. Also, we incorporate the idea of reinforcement learning to actively select both the informative samples and the high-quality annotators, which better explores the trade-off between exploitation and exploration. The experiments clearly demonstrate the efficacy of the proposed model.
Similar papers:
  • Efficient Classifier Training to Minimize False Merges in Electron Microscopy Segmentation [pdf] - Toufiq Parag, Dan C. Ciresan, Alessandro Giusti
  • Introducing Geometry in Active Learning for Image Segmentation [pdf] - Ksenia Konyushkova, Raphael Sznitman, Pascal Fua
  • Point Triangulation Through Polyhedron Collapse Using the l[?] Norm [pdf] - Simon Donne, Bart Goossens, Wilfried Philips
  • Context Aware Active Learning of Activity Recognition Models [pdf] - Mahmudul Hasan, Amit K. Roy-Chowdhury
  • Active Transfer Learning With Zero-Shot Priors: Reusing Past Datasets for Future Tasks [pdf] - Efstratios Gavves, Thomas Mensink, Tatiana Tommasi, Cees G. M. Snoek, Tinne Tuytelaars
Contour Box: Rejecting Object Proposals Without Explicit Closed Contours [pdf]
Cewu Lu, Shu Liu, Jiaya Jia, Chi-Keung Tang

Abstract: Closed contour is an important objectness indicator. We propose a new measure subject to the completeness and tightness constraints, where the optimized closed contour should be tightly bounded within an object proposal. The closed contour measure is defined using closed path integral, and we solve the optimization problem efficiently in polar coordinate system with a global optimum guaranteed. Extensive experiments show that our method can reject a large number of false proposals, and achieve over 6% improvement in object recall at the challenging overlap threshold 0.8 on the VOC 2007 test dataset.
Similar papers:
  • Box Aggregation for Proposal Decimation: Last Mile of Object Detection [pdf] - Shu Liu, Cewu Lu, Jiaya Jia
  • Contour Flow: Middle-Level Motion Estimation by Combining Motion Segmentation and Contour Alignment [pdf] - Huijun Di, Qingxuan Shi, Feng Lv, Ming Qin, Yao Lu
  • Cascaded Sparse Spatial Bins for Efficient and Effective Generic Object Detection [pdf] - David Novotny, Jiri Matas
  • DeepProposal: Hunting Objects by Cascading Deep Convolutional Layers [pdf] - Amir Ghodrati, Ali Diba, Marco Pedersoli, Tinne Tuytelaars, Luc Van Gool
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
Deep Multi-Patch Aggregation Network for Image Style, Aesthetics, and Quality Estimation [pdf]
Xin Lu, Zhe Lin, Xiaohui Shen, Radomir Mech, James Z. Wang

Abstract: This paper investigates problems of image style, aesthetics, and quality estimation, which require fine-grained details from high-resolution images, utilizing deep neural network training approach. Existing deep convolutional neural networks mostly extracted one patch such as a downsized crop from each image as a training example. However, one patch may not always well represent the entire image, which may cause ambiguity during training. We propose a deep multi-patch aggregation network training approach, which allows us to train models using multiple patches generated from one image. We achieve this by constructing multiple, shared columns in the neural network and feeding multiple patches to each of the columns. More importantly, we propose two novel network layers (statistics and sorting) to support aggregation of those patches. The proposed deep multi-patch aggregation network integrates shared feature learning and aggregation function learning into a unified framework. We demonstrate the effectiveness of the deep multi-patch aggregation network on the three problems, i.e., image style recognition, aesthetic quality categorization, and image quality estimation. Our models trained using the proposed networks significantly outperformed the state of the art in all three applications.
Similar papers:
  • Deep Networks for Image Super-Resolution With Sparse Prior [pdf] - Zhaowen Wang, Ding Liu, Jianchao Yang, Wei Han, Thomas Huang
  • Local Convolutional Features With Unsupervised Training for Image Retrieval [pdf] - Mattis Paulin, Matthijs Douze, Zaid Harchaoui, Julien Mairal, Florent Perronin, Cordelia Schmid
  • Unsupervised Visual Representation Learning by Context Prediction [pdf] - Carl Doersch, Abhinav Gupta, Alexei A. Efros
  • Unsupervised Learning of Visual Representations Using Videos [pdf] - Xiaolong Wang, Abhinav Gupta
  • Aggregating Local Deep Features for Image Retrieval [pdf] - Artem Babenko, Victor Lempitsky
Localize Me Anywhere, Anytime: A Multi-Task Point-Retrieval Approach [pdf]
Guoyu Lu, Yan Yan, Li Ren, Jingkuan Song, Nicu Sebe, Chandra Kambhamettu

Abstract: Image-based localization is an essential complement to GPS localization. Current image-based localization methods are based on either 2D-to-3D or 3D-to-2D to find the correspondences, which ignore the real scene geometric attributes. The main contribution of our paper is that we use a 3D model reconstructed by a short video as the query to realize 3D-to-3D localization under a multi-task point retrieval framework. Firstly, the use of a 3D model as the query enables us to efficiently select location candidates. Furthermore, the reconstruction of 3D model exploits the correlation among different images, based on the fact that images captured from different views for SfM share information through matching features. By exploring shared information (matching features) across multiple related tasks (images of the same scene captured from different views), the visual features view-invariance property can be improved in order to get to a higher point retrieval accuracy. More specifically, we use multi-task point retrieval framework to explore the relationship between descriptors and the 3D points, which extracts the discriminant points for more accurate 3D-to-3D correspondences retrieval. We further apply multi-task learning (MTL) retrieval approach on thermal images to prove that our MTL retrieval framework also provides superior performance for the thermal domain. This application is exceptionally helpful to cope with the localization problem in an environment with limited or even
Similar papers:
  • Multi-Label Cross-Modal Retrieval [pdf] - Viresh Ranjan, Nikhil Rasiwasia, C. V. Jawahar
  • Camera Pose Voting for Large-Scale Image-Based Localization [pdf] - Bernhard Zeisl, Torsten Sattler, Marc Pollefeys
  • Predicting Good Features for Image Geo-Localization Using Per-Bundle VLAD [pdf] - Hyo Jin Kim, Enrique Dunn, Jan-Michael Frahm
  • You Are Here: Mimicking the Human Thinking Process in Reading Floor-Plans [pdf] - Hang Chu, Dong Ki Kim, Tsuhan Chen
  • Hyperpoints and Fine Vocabularies for Large-Scale Location Recognition [pdf] - Torsten Sattler, Michal Havlena, Filip Radenovic, Konrad Schindler, Marc Pollefeys
Robust RGB-D Odometry Using Point and Line Features [pdf]
Yan Lu, Dezhen Song

Abstract: Lighting variation and uneven feature distribution are main challenges for indoor RGB-D visual odometry where color information is often combined with depth information. To meet the challenges, we fuse point and line features to form a robust odometry algorithm. Line features are abundant indoors and less sensitive to lighting change than points. We extract 3D points and lines from RGB-D data, analyze their measurement uncertainties, and compute camera motion using maximum likelihood estimation. We prove that fusing points and lines produces smaller motion estimate uncertainty than using either feature type alone. In experiments we compare our method with state-ofthe-art methods including a keypoint-based approach and a dense visual odometry algorithm. Our method outperforms the counterparts under both constant and varying lighting conditions. Specifically, our method achieves an average translational error that is 34.9% smaller than the counterparts, when tested using public datasets.
Similar papers:
  • Model-Based Tracking at 300Hz Using Raw Time-of-Flight Observations [pdf] - Jan Stuhmer, Sebastian Nowozin, Andrew Fitzgibbon, Richard Szeliski, Travis Perry, Sunil Acharya, Daniel Cremers, Jamie Shotton
  • Dual-Feature Warping-Based Motion Model Estimation [pdf] - Shiwei Li, Lu Yuan, Jian Sun, Long Quan
  • Large Displacement 3D Scene Flow With Occlusion Reasoning [pdf] - Andrei Zanfir, Cristian Sminchisescu
  • An Efficient Minimal Solution for Multi-Camera Motion [pdf] - Jonathan Ventura, Clemens Arth, Vincent Lepetit
  • Realtime Edge-Based Visual Odometry for a Monocular Camera [pdf] - Juan Jose Tarrio, Sol Pedre
Simultaneous Local Binary Feature Learning and Encoding for Face Recognition [pdf]
Jiwen Lu, Venice Erin Liong, Jie Zhou

Abstract: In this paper, we propose a simultaneous local binary feature learning and encoding (SLBFLE) method for face recognition. Different from existing hand-crafted face descriptors such as local binary pattern (LBP) and Gabor features which require strong prior knowledge, our SLBFLE is an unsupervised feature learning approach which is automatically learned from raw pixels. Unlike existing binary face descriptors such as the LBP and discriminant face descriptor (DFD) which use a two-stage feature extraction approach, our SLBFLE jointly learns binary codes for local face patches and the codebook for feature encoding so that discriminative information from raw pixels can be simultaneously learned with a one-stage procedure. Experimental results on four widely used face datasets including LFW, YouTube Face (YTF), FERET and PaSC clearly demonstrate the effectiveness of the proposed method. -2 5 4 10 9 8 -3 -2 4 4 5 PDV Training Images Test Image 6 5 1 1 2 -1 -1 0 2 3 4 SLBFLE W D 1 0 0 1 1 1 0 Compact Binary Code Histogram Representation PDV Figure 1. The basic idea of the proposed SLBFLE approach for face representation. For each training face image, we extract pixel difference vectors (PDVs) and jointly learn a discriminative mapping and a dictionary for feature extraction. The mapping is to project each PDV into a low-dimensional binary vector, and the dictionary is used as the codebook for feature local encoding. For each test image, the PDVs are first extracted and
Similar papers:
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • Deep Learning Face Attributes in the Wild [pdf] - Ziwei Liu, Ping Luo, Xiaogang Wang, Xiaoou Tang
  • Selective Encoding for Recognizing Unreliably Localized Faces [pdf] - Ang Li, Vlad Morariu, Larry S. Davis
  • Personalized Age Progression With Aging Dictionary [pdf] - Xiangbo Shu, Jinhui Tang, Hanjiang Lai, Luoqi Liu, Shuicheng Yan
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
Square Localization for Efficient and Accurate Object Detection [pdf]
Cewu Lu, Yongyi Lu, Hao Chen, Chi-Keung Tang

Abstract: The key contribution of this paper is the compact square object localization, which relaxes the exhaustive sliding window from testing all windows of different combinations of aspect ratios. Square object localization is category scalable. By using a binary search strategy, the number of scales to test is further reduced empirically to only O(log(min{H, W })) rounds of sliding CNNs, where H and W are respectively the image height and width. In the training phase, square CNN models and object co-presence priors are learned. In the testing phase, sliding CNN models are applied which produces a set of response maps that can be effectively filtered by the learned co-presence prior to output the final bounding boxes for localizing an object. We performed extensive experimental evaluation on the VOC 2007 and 2012 datasets to demonstrate that while efficient, square localization can output precise bounding boxes to improve the final detection result.
Similar papers:
  • Active Object Localization With Deep Reinforcement Learning [pdf] - Juan C. Caicedo, Svetlana Lazebnik
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • AttentionNet: Aggregating Weak Directions for Accurate Object Detection [pdf] - Donggeun Yoo, Sunggyun Park, Joon-Young Lee, Anthony S. Paek, In So Kweon
  • DeepProposal: Hunting Objects by Cascading Deep Convolutional Layers [pdf] - Amir Ghodrati, Ali Diba, Marco Pedersoli, Tinne Tuytelaars, Luc Van Gool
  • Oriented Object Proposals [pdf] - Shengfeng He, Rynson W.H. Lau
Actionness-Assisted Recognition of Actions [pdf]
Ye Luo, Loong-Fah Cheong, An Tran

Abstract: We elicit from a fundamental definition of action lowlevel attributes that can reveal agency and intentionality. These descriptors are mainly trajectory-based, measuring sudden changes, temporal synchrony, and repetitiveness. The actionness map can be used to localize actions in a way that is generic across action and agent types. Furthermore, it also groups interacting regions into a useful unit of analysis, which is crucial for recognition of actions involving interactions. We then implement an actionness-driven pooling scheme to improve action recognition performance. Experimental results on various datasets show the advantages of our method on action detection and action recognition comparing with other state-of-the-art methods.
Similar papers:
  • Action Localization in Videos Through Context Walk [pdf] - Khurram Soomro, Haroon Idrees, Mubarak Shah
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
  • Action Detection by Implicit Intentional Motion Clustering [pdf] - Wei Chen, Jason J. Corso
Removing Rain From a Single Image via Discriminative Sparse Coding [pdf]
Yu Luo, Yong Xu, Hui Ji

Abstract: Visual distortions on images caused by bad weather conditions can have a negative impact on the performance of many outdoor vision systems. One often seen bad weather is rain which causes significant yet complex local intensity fluctuations in images. The paper aims at developing an effective algorithm to remove visual effects of rain from a single rain image, i.e. separate the rain layer and the derained image layer from a rain image. Built upon a nonlinear generative model of rain image, namely screen blend model, we propose a dictionary learning based algorithm for single image de-raining. The basic idea is to sparsely approximate the patches of two layers by very high discriminative codes over a learned dictionary with strong mutual exclusivity property. Such discriminative sparse codes lead to accurate separation of two layers from their non-linear composite. The experiments show that the proposed method outperforms the existing single image de-raining methods on tested rain images.
Similar papers:
  • Deep Networks for Image Super-Resolution With Sparse Prior [pdf] - Zhaowen Wang, Ding Liu, Jianchao Yang, Wei Han, Thomas Huang
  • Linearization to Nonlinear Learning for Visual Tracking [pdf] - Bo Ma, Hongwei Hu, Jianbing Shen, Yuping Zhang, Fatih Porikli
  • Variational Depth Superresolution Using Example-Based Edge Representations [pdf] - David Ferstl, Matthias Ruther, Horst Bischof
  • Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf] - Yuhui Quan, Yan Huang, Hui Ji
  • Conformal and Low-Rank Sparse Representation for Image Restoration [pdf] - Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng
Hierarchical Convolutional Features for Visual Tracking [pdf]
Chao Ma, Jia-Bin Huang, Xiaokang Yang, Ming-Hsuan Yang

Abstract: Visual object tracking is challenging as target objects often undergo significant appearance changes caused by deformation, abrupt motion, background clutter and occlusion. In this paper, we exploit features extracted from deep convolutional neural networks trained on object recognition datasets to improve tracking accuracy and robustness. The outputs of the last convolutional layers encode the semantic information of targets and such representations are robust to significant appearance variations. However, their spatial resolution is too coarse to precisely localize targets. In contrast, earlier convolutional layers provide more precise localization but are less invariant to appearance changes. We interpret the hierarchies of convolutional layers as a nonlinear counterpart of an image pyramid representation and exploit these multiple levels of abstraction for visual tracking. Specifically, we adaptively learn correlation filters on each convolutional layer to encode the target appearance. We hierarchically infer the maximum response of each layer to locate targets. Extensive experimental results on a largescale benchmark dataset show that the proposed algorithm performs favorably against state-of-the-art methods.
Similar papers:
  • SOWP: Spatially Ordered and Weighted Patch Descriptor for Visual Tracking [pdf] - Han-Ul Kim, Dae-Youn Lee, Jae-Young Sim, Chang-Su Kim
  • Understanding and Diagnosing Visual Tracking Systems [pdf] - Naiyan Wang, Jianping Shi, Dit-Yan Yeung, Jiaya Jia
  • Visual Tracking With Fully Convolutional Networks [pdf] - Lijun Wang, Wanli Ouyang, Xiaogang Wang, Huchuan Lu
  • Multi-Kernel Correlation Filter for Visual Tracking [pdf] - Ming Tang, Jiayi Feng
  • Learning Spatially Regularized Correlation Filters for Visual Tracking [pdf] - Martin Danelljan, Gustav Hager, Fahad Shahbaz Khan, Michael Felsberg
Linearization to Nonlinear Learning for Visual Tracking [pdf]
Bo Ma, Hongwei Hu, Jianbing Shen, Yuping Zhang, Fatih Porikli

Abstract: Due to unavoidable appearance variations caused by occlusion, deformation, and other factors, classifiers for visual tracking are nonlinear as a necessity. Building on the theory of globally linear approximations to nonlinear functions, we introduce an elegant method that jointly learns a nonlinear classifier and a visual dictionary for tracking objects in a semi-supervised sparse coding fashion. This establishes an obvious distinction from conventional sparse coding based discriminative tracking algorithms that usually maintain two-stage learning strategies, i.e., learning a dictionary in an unsupervised way then followed by training a classifier. However, the treating dictionary learning and classifier training as separate stages may not produce both descriptive and discriminative models for objects. By contrast, our method is capable of constructing a dictionary that not only fully reflects the intrinsic manifold structure of the data, but also possesses discriminative power. This paper presents an optimization method to obtain such an optimal dictionary, associated sparse coding, and a classifier in an iterative process. Our experiments on a benchmark show our tracker attains outstanding performance compared with the state-of-the-art algorithms.
Similar papers:
  • Convolutional Sparse Coding for Image Super-Resolution [pdf] - Shuhang Gu, Wangmeng Zuo, Qi Xie, Deyu Meng, Xiangchu Feng, Lei Zhang
  • Discriminative Low-Rank Tracking [pdf] - Yao Sui, Yafei Tang, Li Zhang
  • Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf] - Yuhui Quan, Yan Huang, Hui Ji
  • Adaptively Unified Semi-Supervised Dictionary Learning With Active Points [pdf] - Xiaobo Wang, Xiaojie Guo, Stan Z. Li
  • Conformal and Low-Rank Sparse Representation for Image Restoration [pdf] - Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng
Local Subspace Collaborative Tracking [pdf]
Lin Ma, Xiaoqin Zhang, Weiming Hu, Junliang Xing, Jiwen Lu, Jie Zhou

Abstract: Subspace models have been widely used for appearance based object tracking. Most existing subspace based trackers employ a linear subspace to represent object appearances, which are not accurate enough to model large variations of objects. To address this, this paper presents a local subspace collaborative tracking method for robust visual tracking, where multiple linear and nonlinear subspaces are learned to better model the nonlinear relationship of object appearances. First, we retain a set of key samples and compute a set of local subspaces for each key sample. Then, we construct a hypersphere to represent the local nonlinear subspace for each key sample. The hypersphere of one key sample passes the local key samples and also is tangent to the local linear subspace of the specific key sample. In this way, we are able to represent the nonlinear distribution of the key samples and also approximate the local linear subspace near the specific key sample, so that local distributions of the samples can be represented more accurately. Experimental results on challenging video sequences demonstrate the effectiveness of our method.
Similar papers:
  • Multi-Kernel Correlation Filter for Visual Tracking [pdf] - Ming Tang, Jiayi Feng
  • Multi-View Subspace Clustering [pdf] - Hongchang Gao, Feiping Nie, Xuelong Li, Heng Huang
  • Linearization to Nonlinear Learning for Visual Tracking [pdf] - Bo Ma, Hongwei Hu, Jianbing Shen, Yuping Zhang, Fatih Porikli
  • Discriminative Low-Rank Tracking [pdf] - Yao Sui, Yafei Tang, Li Zhang
  • Multiple Feature Fusion via Weighted Entropy for Visual Tracking [pdf] - Lin Ma, Jiwen Lu, Jianjiang Feng, Jie Zhou
Multimodal Convolutional Neural Networks for Matching Image and Sentence [pdf]
Lin Ma, Zhengdong Lu, Lifeng Shang, Hang Li

Abstract: In this paper, we propose multimodal convolutional neural networks (m-CNNs) for matching image and sentence. Our m-CNN provides an end-to-end framework with convolutional architectures to exploit image representation, word composition, and the matching relations between the two modalities. More specifically, it consists of one image CNN encoding the image content and one matching CNN modeling the joint representation of image and sentence. The matching CNN composes different semantic fragments from words and learns the inter-modal relations between image and the composed fragments at different levels, thus fully exploit the matching relations between image and sentence. Experimental results demonstrate that the proposed m-CNNs can effectively capture the information necessary for image and sentence matching. More specifically, our proposed m-CNNs significantly outperform the state-ofthe-art approaches for bidirectional image and sentence retrieval on the Flickr8K and Flickr30K datasets.
Similar papers:
  • Learning Like a Child: Fast Novel Visual Concept Learning From Sentence Descriptions of Images [pdf] - Junhua Mao, Xu Wei, Yi Yang, Jiang Wang, Zhiheng Huang, Alan L. Yuille
  • 3D Fragment Reassembly Using Integrated Template Guidance and Fracture-Region Matching [pdf] - Kang Zhang, Wuyi Yu, Mary Manhein, Warren Waggenspack, Xin Li
  • Common Subspace for Model and Similarity: Phrase Learning for Caption Generation From Images [pdf] - Yoshitaka Ushiku, Masataka Yamaguchi, Yusuke Mukuta, Tatsuya Harada
  • Guiding the Long-Short Term Memory Model for Image Caption Generation [pdf] - Xu Jia, Efstratios Gavves, Basura Fernando, Tinne Tuytelaars
  • Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books [pdf] - Yukun Zhu, Ryan Kiros, Rich Zemel, Ruslan Salakhutdinov, Raquel Urtasun, Antonio Torralba, Sanja Fidler
Multiple Feature Fusion via Weighted Entropy for Visual Tracking [pdf]
Lin Ma, Jiwen Lu, Jianjiang Feng, Jie Zhou

Abstract: It is desirable to combine multiple feature descriptors to improve the visual tracking performance because different features can provide complementary information to describe objects of interest. However, how to effectively fuse multiple features remains a challenging problem in visual tracking, especially in a data-driven manner. In this paper, we propose a new data-adaptive visual tracking approach by using multiple feature fusion via weighted entropy. Unlike existing visual trackers which simply concatenate multiple feature vectors together for object representation, we employ the weighted entropy to evaluate the dissimilarity between the object state and the background state, and seek the optimal feature combination by minimizing the weighted entropy, so that more complementary information can be exploited for object representation. Experimental results demonstrate the effectiveness of our approach in tackling various challenges for visual object tracking.
Similar papers:
  • Multi-Kernel Correlation Filter for Visual Tracking [pdf] - Ming Tang, Jiayi Feng
  • Linearization to Nonlinear Learning for Visual Tracking [pdf] - Bo Ma, Hongwei Hu, Jianbing Shen, Yuping Zhang, Fatih Porikli
  • SOWP: Spatially Ordered and Weighted Patch Descriptor for Visual Tracking [pdf] - Han-Ul Kim, Dae-Youn Lee, Jae-Young Sim, Chang-Su Kim
  • Discriminative Low-Rank Tracking [pdf] - Yao Sui, Yafei Tang, Li Zhang
  • Local Subspace Collaborative Tracking [pdf] - Lin Ma, Xiaoqin Zhang, Weiming Hu, Junliang Xing, Jiwen Lu, Jie Zhou
Ask Your Neurons: A Neural-Based Approach to Answering Questions About Images [pdf]
Mateusz Malinowski, Marcus Rohrbach, Mario Fritz

Abstract: We address a question answering task on real-world images that is set up as a Visual Turing Test. By combining latest advances in image representation and natural language processing, we propose Neural-Image-QA, an endto-end formulation to this problem for which all parts are trained jointly. In contrast to previous efforts, we are facing a multi-modal problem where the language output (answer) is conditioned on visual and natural language input (image and question). Our approach Neural-Image-QA doubles the performance of the previous best approach on this problem. We provide additional insights into the problem by analyzing how much information is contained only in the language part for which we provide a new human baseline. To study human consensus, which is related to the ambiguities inherent in this challenging task, we propose two novel metrics and collect additional answers which extends the original DAQUAR dataset to DAQUAR-Consensus. CNN What is behind the table ? LSTM LSTM LSTM LSTM LSTM LSTM LSTM LSTM chairs window Figure 1. Our approach Neural-Image-QA to question answering with a Recurrent Neural Network using Long Short Term Memory (LSTM). To answer a question about an image, we feed in both, the image (CNN features) and the question (green boxes) into the LSTM. After the (variable length) question is encoded, we generate the answers (multiple words, orange boxes). During the answer generation phase the previously predicted answers are fed
Similar papers:
  • LEWIS: Latent Embeddings for Word Images and their Semantics [pdf] - Albert Gordo, Jon Almazan, Naila Murray, Florent Perronin
  • Guiding the Long-Short Term Memory Model for Image Caption Generation [pdf] - Xu Jia, Efstratios Gavves, Basura Fernando, Tinne Tuytelaars
  • Sequence to Sequence - Video to Text [pdf] - Subhashini Venugopalan, Marcus Rohrbach, Jeffrey Donahue, Raymond Mooney, Trevor Darrell, Kate Saenko
  • Visual Madlibs: Fill in the Blank Description Generation and Question Answering [pdf] - Licheng Yu, Eunbyung Park, Alexander C. Berg, Tamara L. Berg
  • VQA: Visual Question Answering [pdf] - Stanislaw Antol, Aishwarya Agrawal, Jiasen Lu, Margaret Mitchell, Dhruv Batra, C. Lawrence Zitnick, Devi Parikh
FaceDirector: Continuous Control of Facial Performance in Video [pdf]
Charles Malleson, Jean-Charles Bazin, Oliver Wang, Derek Bradley, Thabo Beeler, Adrian Hilton, Alexander Sorkine-Hornung

Abstract: We present a method to continuously blend between multiple facial performances of an actor, which can contain different facial expressions or emotional states. As an example, given sad and angry video takes of a scene, our method empowers a movie director to specify arbitrary weighted combinations and smooth transitions between the two takes in post-production. Our contributions include (1) a robust nonlinear audio-visual synchronization technique that exploits complementary properties of audio and visual cues to automatically determine robust, dense spatio-temporal correspondences between takes, and (2) a seamless facial blending approach that provides the director full control to interpolate timing, facial expression, and local appearance, in order to generate novel performances after filming. In contrast to most previous works, our approach operates entirely in image space, avoiding the need of 3D facial reconstruction. We demonstrate that our method can synthesize visually believable performances with applications in emotion transition, performance correction, and timing control
Similar papers:
  • Listening With Your Eyes: Towards a Practical Visual Speech Recognition System Using Deep Boltzmann Machines [pdf] - Chao Sui, Mohammed Bennamoun, Roberto Togneri
  • Pairwise Conditional Random Forests for Facial Expression Recognition [pdf] - Arnaud Dapogny, Kevin Bailly, Severine Dubuisson
  • Face Flow [pdf] - Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou
  • From Emotions to Action Units With Hidden and Semi-Hidden-Task Learning [pdf] - Adria Ruiz, Joost Van de Weijer, Xavier Binefa
  • Joint Fine-Tuning in Deep Neural Networks for Facial Expression Recognition [pdf] - Heechul Jung, Sihaeng Lee, Junho Yim, Sunjeong Park, Junmo Kim
Learning Informative Edge Maps for Indoor Scene Layout Prediction [pdf]
Arun Mallya, Svetlana Lazebnik

Abstract: In this paper, we introduce new edge-based features for the task of recovering the 3D layout of an indoor scene from a single image. Indoor scenes have certain edges that are very informative about the spatial layout of the room, namely, the edges formed by the pairwise intersections of room faces (two walls, wall and ceiling, wall and floor). In contrast with previous approaches that rely on area-based features like geometric context and orientation maps, our method attempts to directly detect these informative edges. We learn to predict informative edge probability maps using two recent methods that exploit local and global context, respectively: structured edge detection forests, and a fully convolutional network for pixelwise labeling. We show that the fully convolutional network is quite successful at predicting the informative edges even when they lack contrast or are occluded, and that the accuracy can be further improved by training the network to jointly predict the edges and the geometric context. Using features derived from the informative edge maps, we learn a maximum margin structured classifier that achieves state-of-the-art performance on layout prediction.
Similar papers:
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • Structured Indoor Modeling [pdf] - Satoshi Ikehata, Hang Yang, Yasutaka Furukawa
  • Realtime Edge-Based Visual Odometry for a Monocular Camera [pdf] - Juan Jose Tarrio, Sol Pedre
  • Lost Shopping! Monocular Localization in Large Indoor Spaces [pdf] - Shenlong Wang, Sanja Fidler, Raquel Urtasun
  • Holistically-Nested Edge Detection [pdf] - Saining Xie, Zhuowen Tu
Learning Like a Child: Fast Novel Visual Concept Learning From Sentence Descriptions of Images [pdf]
Junhua Mao, Xu Wei, Yi Yang, Jiang Wang, Zhiheng Huang, Alan L. Yuille

Abstract: Model-base a group of people playing a game of soccer Train using quidditch images with sentence descriptions In this paper, we address the task of learning novel visual concepts, and their interactions with other concepts, from a few images with sentence descriptions. Using linguistic context and visual features, our method is able to efficiently hypothesize the semantic meaning of new words and add them to its word dictionary so that they can be used to describe images which contain these novel concepts. Our method has an image captioning module based on [38] with several improvements. In particular, we propose a transposed weight sharing scheme, which not only improves performance on image captioning, but also makes the model more suitable for the novel concept learning task. We propose methods to prevent overfitting the new concepts. In addition, three novel concept datasets are constructed for this new task, and are publicly available on the project page. In the experiments, we show that our method effectively learns novel visual concepts from a few examples without disturbing the previously learned concepts. The project page is: www.stat.ucla.edu/junhua. mao/projects/child_learning.html.
Similar papers:
  • Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books [pdf] - Yukun Zhu, Ryan Kiros, Rich Zemel, Ruslan Salakhutdinov, Raquel Urtasun, Antonio Torralba, Sanja Fidler
  • Guiding the Long-Short Term Memory Model for Image Caption Generation [pdf] - Xu Jia, Efstratios Gavves, Basura Fernando, Tinne Tuytelaars
  • Selecting Relevant Web Trained Concepts for Automated Event Retrieval [pdf] - Bharat Singh, Xintong Han, Zhe Wu, Vlad I. Morariu, Larry S. Davis
  • LEWIS: Latent Embeddings for Word Images and their Semantics [pdf] - Albert Gordo, Jon Almazan, Naila Murray, Florent Perronin
  • Automatic Concept Discovery From Parallel Text and Visual Corpora [pdf] - Chen Sun, Chuang Gan, Ram Nevatia
Thin Structure Estimation With Curvature Regularization [pdf]
Dmitrii Marin, Yuchen Zhong, Maria Drangova, Yuri Boykov

Abstract: Many applications in vision require estimation of thin structures such as boundary edges, surfaces, roads, blood vessels, neurons, etc. Unlike most previous approaches, we simultaneously detect and delineate thin structures with sub-pixel localization and real-valued orientation estimation. This is an ill-posed problem that requires regularization. We propose an objective function combining detection likelihoods with a prior minimizing curvature of the centerlines or surfaces. Unlike simple block-coordinate descent, we develop a novel algorithm that is able to perform joint optimization of location and detection variables more effectively. Our lower bound optimization algorithm applies to quadratic or absolute curvature. The proposed early vision framework is sufficiently general and it can be used in many higher-level applications. We illustrate the advantage of our approach on a range of 2D and 3D examples.
Similar papers:
  • Realtime Edge-Based Visual Odometry for a Monocular Camera [pdf] - Juan Jose Tarrio, Sol Pedre
  • A Multiscale Variable-Grouping Framework for MRF Energy Minimization [pdf] - Omer Meir, Meirav Galun, Stav Yagev, Ronen Basri, Irad Yavneh
  • Volumetric Bias in Segmentation and Reconstruction: Secrets and Solutions [pdf] - Yuri Boykov, Hossam Isack, Carl Olsson, Ismail Ben Ayed
  • On the Visibility of Point Clouds [pdf] - Sagi Katz, Ayellet Tal
  • Component-Wise Modeling of Articulated Objects [pdf] - Valsamis Ntouskos, Marta Sanzari, Bruno Cafaro, Federico Nardi, Fabrizio Natola, Fiora Pirri, Manuel Ruiz
3D Time-Lapse Reconstruction From Internet Photos [pdf]
Ricardo Martin-Brualla, David Gallup, Steven M. Seitz

Abstract: Given an Internet photo collection of a landmark, we compute a 3D time-lapse video sequence where a virtual camera moves continuously in time and space. While previous work assumed a static camera, the addition of camera motion during the time-lapse creates a very compelling impression of parallax. Achieving this goal, however, requires addressing multiple technical challenges, including solving for time-varying depth maps, regularizing 3D point color profiles over time, and reconstructing high quality, hole-free images at every frame from the projected profiles. Our results show photorealistic time-lapses of skylines and natural scenes over many years, with dramatic parallax effects.
Similar papers:
  • Variational PatchMatch MultiView Reconstruction and Refinement [pdf] - Philipp Heise, Brian Jensen, Sebastian Klose, Alois Knoll
  • Sparse Dynamic 3D Reconstruction From Unsynchronized Videos [pdf] - Enliang Zheng, Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm
  • What Makes Tom Hanks Look Like Tom Hanks [pdf] - Supasorn Suwajanakorn, Steven M. Seitz, Ira Kemelmacher-Shlizerman
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
  • Synthesizing Illumination Mosaics From Internet Photo-Collections [pdf] - Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm
PQTable: Fast Exact Asymmetric Distance Neighbor Search for Product Quantization Using Hash Tables [pdf]
Yusuke Matsui, Toshihiko Yamasaki, Kiyoharu Aizawa

Abstract: We propose the product quantization table (PQTable), a product quantization-based hash table that is fast and requires neither parameter tuning nor training steps. The PQTable produces exactly the same results as a linear PQ search, and is 102 to 105 times faster when tested on the SIFT1B data. In addition, although state-of-the-art performance can be achieved by previous inverted-indexing-based approaches, such methods do require manually designed parameter setting and much training, whereas our method is free from them. Therefore, PQTable offers a practical and useful solution for real-world problems.
Similar papers:
  • Two Birds, One Stone: Jointly Learning Binary Code for Large-Scale Face Image Retrieval and Attributes Prediction [pdf] - Yan Li, Ruiping Wang, Haomiao Liu, Huajie Jiang, Shiguang Shan, Xilin Chen
  • Projection Bank: From High-Dimensional Data to Medium-Length Binary Codes [pdf] - Li Liu, Mengyang Yu, Ling Shao
  • Alternating Co-Quantization for Cross-Modal Hashing [pdf] - Go Irie, Hiroyuki Arai, Yukinobu Taniguchi
  • Multi-View Complementary Hash Tables for Nearest Neighbor Search [pdf] - Xianglong Liu, Lei Huang, Cheng Deng, Jiwen Lu, Bo Lang
  • Learning Binary Codes for Maximum Inner Product Search [pdf] - Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen
Enhancing Road Maps by Parsing Aerial Images Around the World [pdf]
Gellert Mattyus, Shenlong Wang, Sanja Fidler, Raquel Urtasun

Abstract: In recent years, contextual models that exploit maps have been shown to be very effective for many recognition and localization tasks. In this paper we propose to exploit aerial images in order to enhance freely available world maps. Towards this goal, we make use of OpenStreetMap and formulate the problem as the one of inference in a Markov random field parameterized in terms of the location of the road-segment centerlines as well as their width. This parameterization enables very efficient inference and returns only topologically correct roads. In particular, we can segment all OSM roads in the whole world in a single day using a small cluster of 10 computers. Importantly, our approach generalizes very well; it can be trained using only 1.5 km2 aerial imagery and produce very accurate results in any location across the globe. We demonstrate the effectiveness of our approach outperforming the state-of-the-art in two new benchmarks that we collect. We then show how our enhanced maps are beneficial for semantic segmentation of ground images.
Similar papers:
  • Cutting Edge: Soft Correspondences in Multimodal Scene Parsing [pdf] - Sarah Taghavi Namin, Mohammad Najafi, Mathieu Salzmann, Lars Petersson
  • Car That Knows Before You Do: Anticipating Maneuvers via Learning Temporal Driving Models [pdf] - Ashesh Jain, Hema S. Koppula, Bharad Raghavan, Shane Soh, Ashutosh Saxena
  • Semantically-Aware Aerial Reconstruction From Multi-Modal Data [pdf] - Randi Cabezas, Julian Straub, John W. Fisher III
  • Projection Onto the Manifold of Elongated Structures for Accurate Extraction [pdf] - Amos Sironi, Vincent Lepetit, Pascal Fua
  • Wide-Area Image Geolocalization With Aerial Reference Imagery [pdf] - Scott Workman, Richard Souvenir, Nathan Jacobs
BubbLeNet: Foveated Imaging for Visual Discovery [pdf]
Kevin Matzen, Noah Snavely

Abstract: We propose a new method for turning an Internet-scale corpus of categorized images into a small set of humaninterpretable discriminative visual elements using powerful tools based on deep learning. A key challenge with deep learning methods is generating human interpretable models. To address this, we propose a new technique that uses bubble imagesimages where most of the content has been obscuredto identify spatially localized, discriminative content in each image. By modifying the model training procedure to use both the source imagery and these bubble images, we can arrive at final models which retain much of the original classification performance, but are much more amenable to identifying interpretable visual elements. We apply our algorithm to a wide variety of datasets, including two new Internet-scale datasets of people and places, and show applications to visual mining and discovery. Our method is simple, scalable, and produces visual elements that are highly representative compared to prior work.
Similar papers:
  • Dense Optical Flow Prediction From a Static Image [pdf] - Jacob Walker, Abhinav Gupta, Martial Hebert
  • Single Image 3D Without a Single 3D Image [pdf] - David F. Fouhey, Wajahat Hussain, Abhinav Gupta, Martial Hebert
  • Unsupervised Learning of Visual Representations Using Videos [pdf] - Xiaolong Wang, Abhinav Gupta
  • Harvesting Discriminative Meta Objects With Deep CNN Features for Scene Classification [pdf] - Ruobing Wu, Baoyuan Wang, Wenping Wang, Yizhou Yu
  • Unsupervised Visual Representation Learning by Context Prediction [pdf] - Carl Doersch, Abhinav Gupta, Alexei A. Efros
Improving Image Restoration With Soft-Rounding [pdf]
Xing Mei, Honggang Qi, Bao-Gang Hu, Siwei Lyu

Abstract: Several important classes of images such as text, barcode and pattern images have the property that pixels can only take a distinct subset of values. This knowledge can benefit the restoration of such images, but it has not been widely considered in current restoration methods. In this work, we describe an effective and efficient approach to incorporate the knowledge of distinct pixel values of the pristine images into the general regularized least squares restoration framework. We introduce a new regularizer that attains zero at the designated pixel values and becomes a quadratic penalty function in the intervals between them. When incorporated into the regularized least squares restoration framework, this regularizer leads to a simple and efficient step that resembles and extends the rounding operation, which we term as soft-rounding. We apply the soft-rounding enhanced solution to the restoration of binary text/barcode images and pattern images with multiple distinct pixel values. Experimental results show that soft-rounding enhanced restoration methods achieve significant improvement in both visual quality and quantitative measures (PSNR and SSIM). Furthermore, we show that this regularizer can also benefit the restoration of general natural images.
Similar papers:
  • FASText: Efficient Unconstrained Scene Text Detector [pdf] - Michal Busta, Lukas Neumann, Jiri Matas
  • Class-Specific Image Deblurring [pdf] - Saeed Anwar, Cong Phuoc Huynh, Fatih Porikli
  • Compression Artifacts Reduction by a Deep Convolutional Network [pdf] - Chao Dong, Yubin Deng, Chen Change Loy, Xiaoou Tang
  • Video Restoration Against Yin-Yang Phasing [pdf] - Xiaolin Wu, Zhenhao Li, Xiaowei Deng
  • Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf] - Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
A Multiscale Variable-Grouping Framework for MRF Energy Minimization [pdf]
Omer Meir, Meirav Galun, Stav Yagev, Ronen Basri, Irad Yavneh

Abstract: We present a multiscale approach for minimizing the energy associated with Markov Random Fields (MRFs) with energy functions that include arbitrary pairwise potentials. The MRF is represented on a hierarchy of successively coarser scales, where the problem on each scale is itself an MRF with suitably defined potentials. These representations are used to construct an efficient multiscale algorithm that seeks a minimal-energy solution to the original problem. The algorithm is iterative and features a bidirectional crosstalk between fine and coarse representations. We use consistency criteria to guarantee that the energy is nonincreasing throughout the iterative process. The algorithm is evaluated on real-world datasets, achieving competitive performance in relatively short run-times.
Similar papers:
  • Probabilistic Label Relation Graphs With Ising Models [pdf] - Nan Ding, Jia Deng, Kevin P. Murphy, Hartmut Neven
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
  • Inferring M-Best Diverse Labelings in a Single One [pdf] - Alexander Kirillov, Bogdan Savchynskyy, Dmitrij Schlesinger, Dmitry Vetrov, Carsten Rother
  • Volumetric Bias in Segmentation and Reconstruction: Secrets and Solutions [pdf] - Yuri Boykov, Hossam Isack, Carl Olsson, Ismail Ben Ayed
  • Higher-Order Inference for Multi-Class Log-Supermodular Models [pdf] - Jian Zhang, Josip Djolonga, Andreas Krause
Extraction of Virtual Baselines From Distorted Document Images Using Curvilinear Projection [pdf]
Gaofeng Meng, Zuming Huang, Yonghong Song, Shiming Xiang, Chunhong Pan

Abstract: The baselines of a document page are a set of virtual horizontal and parallel lines, to which the printed contents of document, e.g., text lines, tables or inserted photos, are aligned. Accurate baseline extraction is of great importance in the geometric correction of curved document images. In this paper, we propose an efficient method for accurate extraction of these virtual visual cues from a curved document image. Our method comes from two basic observations that the baselines of documents do not intersect with each other and that within a narrow strip, the baselines can be well approximated by linear segments. Based upon these observations, we propose a curvilinear projection based method and model the estimation of curved baselines as a constrained sequential optimization problem. A dynamic programming algorithm is then developed to efficiently solve the problem. The proposed method can extract the complete baselines through each pixel of document images in a high accuracy. It is also scripts insensitive and highly robust to image noises, non-textual objects, image resolutions and image quality degradation like blurring and non-uniform illumination. Extensive experiments on a number of captured document images demonstrate the effectiveness of the proposed method.
Similar papers:
  • Fast Orthogonal Projection Based on Kronecker Product [pdf] - Xu Zhang, Felix X. Yu, Ruiqi Guo, Sanjiv Kumar, Shengjin Wang, Shi-Fu Chang
  • Learning Common Sense Through Visual Abstraction [pdf] - Ramakrishna Vedantam, Xiao Lin, Tanmay Batra, C. Lawrence Zitnick, Devi Parikh
  • Detection and Segmentation of 2D Curved Reflection Symmetric Structures [pdf] - Ching L. Teo, Cornelia Fermuller, Yiannis Aloimonos
  • FASText: Efficient Unconstrained Scene Text Detector [pdf] - Michal Busta, Lukas Neumann, Jiri Matas
  • Text Flow: A Unified Text Detection System in Natural Scene Images [pdf] - Shangxuan Tian, Yifeng Pan, Chang Huang, Shijian Lu, Kai Yu, Chew Lim Tan
Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf]
Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner

Abstract: We present a gonio-plenoptic imaging system that realizes a single-shot shape measurement for specular surfaces. The system is comprised of a collimated illumination source and a plenoptic camera. Unlike a conventional plenoptic camera, our system captures the BRDF variation of the object surface in a single image in addition to the light field information from the scene, which allows us to recover very fine 3D structures of the surface. The shape of the surface is reconstructed based on the reflectance property of the material rather than the parallax between different views. Since only a single-shot is required to reconstruct the whole surface, our system is able to capture dynamic surface deformation in a video mode. We also describe a novel calibration technique that maps the light field viewing directions from the object space to subpixels on the sensor plane. The proposed system is evaluated using a concave mirror with known curvature, and is compared to a parabolic mirror scanning system as well as a multi-illumination photometric stereo approach based on simulations and experiments.
Similar papers:
  • Reflection Modeling for Passive Stereo [pdf] - Rahul Nair, Andrew Fitzgibbon, Daniel Kondermann, Carsten Rother
  • Polarized 3D: High-Quality Depth Sensing With Polarization Cues [pdf] - Achuta Kadambi, Vage Taamazyan, Boxin Shi, Ramesh Raskar
  • Photogeometric Scene Flow for High-Detail Dynamic 3D Reconstruction [pdf] - Paulo F. U. Gotardo, Tomas Simon, Yaser Sheikh, Iain Matthews
  • Photometric Stereo in a Scattering Medium [pdf] - Zak Murez, Tali Treibitz, Ravi Ramamoorthi, David Kriegman
  • Photometric Stereo With Small Angular Variations [pdf] - Jian Wang, Yasuyuki Matsushita, Boxin Shi, Aswin C. Sankaranarayanan
Robust Model-Based 3D Head Pose Estimation [pdf]
Gregory P. Meyer, Shalini Gupta, Iuri Frosio, Dikpal Reddy, Jan Kautz

Abstract: We introduce a method for accurate three dimensional head pose estimation using a commodity depth camera. We perform pose estimation by registering a morphable face model to the measured depth data, using a combination of particle swarm optimization (PSO) and the iterative closest point (ICP) algorithm, which minimizes a cost function that includes a 3D registration and a 2D overlap term. The pose is estimated on the fly without requiring an explicit initialization or training phase. Our method handles large pose angles and partial occlusions by dynamically adapting to the reliable visible parts of the face. It is robust and generalizes to different depth sensors without modification. On the Biwi Kinect dataset, we achieve best-in-class performance, with average angular errors of 2.1, 2.1 and 2.4 degrees for yaw, pitch, and roll, respectively, and an average translational error of 5.9 mm, while running at 6 fps on a graphics processing unit.
Similar papers:
  • Rendering of Eyes for Eye-Shape Registration and Gaze Estimation [pdf] - Erroll Wood, Tadas Baltrusaitis, Xucong Zhang, Yusuke Sugano, Peter Robinson, Andreas Bulling
  • Person Recognition in Personal Photo Collections [pdf] - Seong Joon Oh, Rodrigo Benenson, Mario Fritz, Bernt Schiele
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
  • Uncovering Interactions and Interactors: Joint Estimation of Head, Body Orientation and F-Formations From Surveillance Videos [pdf] - Elisa Ricci, Jagannadan Varadarajan, Ramanathan Subramanian, Samuel Rota Bulo, Narendra Ahuja, Oswald Lanz
  • Fast and Accurate Head Pose Estimation via Random Projection Forests [pdf] - Donghoon Lee, Ming-Hsuan Yang, Songhwai Oh
Im2Calories: Towards an Automated Mobile Vision Food Diary [pdf]
Austin Meyers, Nick Johnston, Vivek Rathod, Anoop Korattikara, Alex Gorban, Nathan Silberman, Sergio Guadarrama, George Papandreou, Jonathan Huang, Kevin P. Murphy

Abstract: ies: towards an automated mobile vision food diary Austin Myers, Nick Johnston, Vivek Rathod, Anoop Korattikara, Alex Gorban Nathan Silberman, Sergio Guadarrama, George Papandreou, Jonathan Huang, Kevin Murphy amyers@umd.edu, (nickj, rathodv, kbanoop, gorban)@google.com (nsilberman, sguada, gpapan, jonathanhuang, kpmurphy)@google.com We present a system which can recognize the contents of your meal from a single image, and then predict its nutritional contents, such as calories. The simplest version assumes that the user is eating at a restaurant for which we know the menu. In this case, we can collect images offline to train a multi-label classifier. At run time, we apply the classifier (running on your phone) to predict which foods are present in your meal, and we lookup the corresponding nutritional facts. We apply this method to a new dataset of images from 23 different restaurants, using a CNN-based classifier, significantly outperforming previous work. The more challenging setting works outside of restaurants. In this case, we need to estimate the size of the foods, as well as their labels. This requires solving segmentation and depth / volume estimation from a single image. We present CNN-based approaches to these problems, with promising preliminary results.
Similar papers:
  • Love Thy Neighbors: Image Annotation by Exploiting Image Metadata [pdf] - Justin Johnson, Lamberto Ballan, Li Fei-Fei
  • Where to Buy It: Matching Street Clothing Photos in Online Shops [pdf] - M. Hadi Kiapour, Xufeng Han, Svetlana Lazebnik, Alexander C. Berg, Tamara L. Berg
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
  • Learning Visual Clothing Style With Heterogeneous Dyadic Co-Occurrences [pdf] - Andreas Veit, Balazs Kovacs, Sean Bell, Julian McAuley, Kavita Bala, Serge Belongie
  • Learning Concept Embeddings With Combined Human-Machine Expertise [pdf] - Michael Wilber, Iljung S. Kwak, David Kriegman, Serge Belongie
Learning Nonlinear Spectral Filters for Color Image Reconstruction [pdf]
Michael Moeller, Julia Diebold, Guy Gilboa, Daniel Cremers

Abstract: This paper presents the idea of learning optimal filters for color image reconstruction based on a novel concept of nonlinear spectral image decompositions recently proposed by Guy Gilboa. We use a multiscale image decomposition approach based on total variation regularization and Bregman iterations to represent the input data as the sum of image layers containing features at different scales. Filtered images can be obtained by weighted linear combinations of the different frequency layers. We introduce the idea of learning optimal filters for the task of image denoising, and propose the idea of mixing high frequency components of different color channels. Our numerical experiments demonstrate that learning the optimal weights can significantly improve the results in comparison to the standard variational approach, and achieves state-of-theart image denoising results.
Similar papers:
  • An Efficient Statistical Method for Image Noise Level Estimation [pdf] - Guangyong Chen, Fengyuan Zhu, Pheng Ann Heng
  • Low-Rank Tensor Approximation With Laplacian Scale Mixture Modeling for Multiframe Image Denoising [pdf] - Weisheng Dong, Guangyu Li, Guangming Shi, Xin Li, Yi Ma
  • External Patch Prior Guided Internal Clustering for Image Denoising [pdf] - Fei Chen, Lei Zhang, Huimin Yu
  • Patch Group Based Nonlocal Self-Similarity Prior Learning for Image Denoising [pdf] - Jun Xu, Lei Zhang, Wangmeng Zuo, David Zhang, Xiangchu Feng
  • Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf] - Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
Spatial Semantic Regularisation for Large Scale Object Detection [pdf]
Damian Mrowca, Marcus Rohrbach, Judy Hoffman, Ronghang Hu, Kate Saenko, Trevor Darrell

Abstract: Large scale object detection with thousands of classes introduces the problem of many contradicting false positive detections, which have to be suppressed. Class-independent non-maximum suppression has traditionally been used for this step, but it does not scale well as the number of classes grows. Traditional non-maximum suppression does not consider label- and instance-level relationships nor does it allow an exploitation of the spatial layout of detection proposals. We propose a new multi-class spatial semantic regularisation method based on affinity propagation clustering [8, 22], which simultaneously optimises across all categories and all proposed locations in the image, to improve both the localisation and categorisation of selected detection proposals. Constraints are shared across the labels through the semantic WordNet hierarchy. Our approach proves to be especially useful in large scale settings with thousands of classes, where spatial and semantic interactions are very frequent and only weakly supervised detectors can be built due to a lack of bounding box annotations. Detection experiments are conducted on the ImageNet and COCO dataset, and in settings with thousands of detected categories. Our method provides a significant precision improvement by reducing false positives, while simultaneously improving the recall.
Similar papers:
  • BoxSup: Exploiting Bounding Boxes to Supervise Convolutional Networks for Semantic Segmentation [pdf] - Jifeng Dai, Kaiming He, Jian Sun
  • AttentionNet: Aggregating Weak Directions for Accurate Object Detection [pdf] - Donggeun Yoo, Sunggyun Park, Joon-Young Lee, Anthony S. Paek, In So Kweon
  • Semi-Supervised Zero-Shot Classification With Label Representation Learning [pdf] - Xin Li, Yuhong Guo, Dale Schuurmans
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • Boosting Object Proposals: From Pascal to COCO [pdf] - Jordi Pont-Tuset, Luc Van Gool
An NMF Perspective on Binary Hashing [pdf]
Lopamudra Mukherjee, Sathya N. Ravi, Vamsi K. Ithapu, Tyler Holmes, Vikas Singh

Abstract: The pervasiveness of massive data repositories has led to much interest in efficient methods for indexing, search, and retrieval. For image data, a rapidly developing body of work for these applications shows impressive performance with methods that broadly fall under the umbrella term of Binary Hashing. Given a distance matrix, a binary hashing algorithm solves for a binary code for the given set of examples, whose Hamming distance nicely approximates the original distances. The formulation is non-convex so existing solutions adopt spectral relaxations or perform coordinate descent (or quantization) on a surrogate objective that is numerically more tractable. In this paper, we first derive an Augmented Lagrangian approach to optimize the standard binary Hashing objective (i.e., maintain fidelity with a given distance matrix). With appropriate step sizes, we find that this scheme already yields results that match or substantially outperform state of the art methods on most benchmarks used in the literature. Then, to allow the model to scale to large datasets, we obtain an interesting reformulation of the binary hashing objective as a non-negative matrix factorization. Later, this leads to a simple multiplicative updates algorithm whose parallelization properties are exploited to obtain a fast GPU based implementation. We give a probabilistic analysis of our initialization scheme and present a range of experiments to show that the method is simple to implement and competes
Similar papers:
  • Adaptive Hashing for Fast Similarity Search [pdf] - Fatih Cakir, Stan Sclaroff
  • Top Rank Supervised Binary Coding for Visual Search [pdf] - Dongjin Song, Wei Liu, Rongrong Ji, David A. Meyer, John R. Smith
  • Learning Binary Codes for Maximum Inner Product Search [pdf] - Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen
  • kNN Hashing With Factorized Neighborhood Representation [pdf] - Kun Ding, Chunlei Huo, Bin Fan, Chunhong Pan
  • Alternating Co-Quantization for Cross-Modal Hashing [pdf] - Go Irie, Hiroyuki Arai, Yukinobu Taniguchi
Building Dynamic Cloud Maps From the Ground Up [pdf]
Calvin Murdock, Nathan Jacobs, Robert Pless

Abstract: Satellite imagery of cloud cover is extremely important for understanding and predicting weather. We demonstrate how this imagery can be constructed from the ground up without requiring expensive geo-stationary satellites. This is accomplished through a novel approach to approximate continental-scale cloud maps using only ground-level imagery from publicly-available webcams. We collected a years worth of satellite data and simultaneously-captured, geo-located outdoor webcam images from 4388 sparsely distributed cameras across the continental USA. The satellite data is used to train a dynamic model of cloud motion alongside 4388 regression models (one for each camera) to relate ground-level webcam data to the satellite data at the cameras location. This novel application of largescale computer vision to meteorology and remote sensing is enabled by a smoothed, hierarchically-regularized dynamic texture model whose system dynamics are driven to remain consistent with measurements from the geo-located webcams. We show that our hierarchical model is better able to incorporate sparse webcam measurements resulting in more accurate cloud maps in comparison to a standard dynamic textures implementation. Finally, we demonstrate that our model can be successfully applied to other natural image sequences from the DynTex database, suggesting a broader applicability of our method.
Similar papers:
  • Sparse Dynamic 3D Reconstruction From Unsynchronized Videos [pdf] - Enliang Zheng, Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm
  • Hot or Not: Exploring Correlations Between Appearance and Temperature [pdf] - Daniel Glasner, Pascal Fua, Todd Zickler, Lihi Zelnik-Manor
  • Hyperspectral Compressive Sensing Using Manifold-Structured Sparsity Prior [pdf] - Lei Zhang, Wei Wei, Yanning Zhang, Fei Li, Chunhua Shen, Qinfeng Shi
  • Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf] - Yuhui Quan, Yan Huang, Hui Ji
  • General Dynamic Scene Reconstruction From Multiple View Video [pdf] - Armin Mustafa, Hansung Kim, Jean-Yves Guillemaut, Adrian Hilton
Semantic Component Analysis [pdf]
Calvin Murdock, Fernando De la Torre

Abstract:  Unsupervised and weakly-supervised visual learning in large image collections are critical in order to avoid the time-consuming and error-prone process of manual labeling. Standard approaches rely on methods like multipleinstance learning or graphical models, which can be computationally intensive and sensitive to initialization. On the other hand, simpler component analysis or clustering methods usually cannot achieve meaningful invariances or semantic interpretability. To address the issues of previous work, we present a simple but effective method called Semantic Component Analysis (SCA), which provides a decomposition of images into semantic components. Unsupervised SCA decomposes additive image representations into spatially-meaningful visual components that naturally correspond to object categories. Using an overcomplete representation that allows for rich instance-level constraints and spatial priors, SCA gives improved results and more interpretable components in comparison to traditional matrix factorization techniques. If weakly-supervised information is available in the form of image-level tags, SCA factorizes a set of images into semantic groups of superpixels. We also provide qualitative connections to traditional methods for component analysis (e.g. Grassmann averages, PCA, and NMF). The effectiveness of our approach is validated through synthetic data and on the MSRC2 and Sift Flow datasets, demonstrating competitive results in unsupervised and weakly-superv
Similar papers:
  • Segment-Phrase Table for Semantic Segmentation, Visual Entailment and Paraphrasing [pdf] - Hamid Izadinia, Fereshteh Sadeghi, Santosh K. Divvala, Hannaneh Hajishirzi, Yejin Choi, Ali Farhadi
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • Semantically-Aware Aerial Reconstruction From Multi-Modal Data [pdf] - Randi Cabezas, Julian Straub, John W. Fisher III
  • Semantic Segmentation of RGBD Images With Mutex Constraints [pdf] - Zhuo Deng, Sinisa Todorovic, Longin Jan Latecki
  • Constrained Convolutional Neural Networks for Weakly Supervised Segmentation [pdf] - Deepak Pathak, Philipp Krahenbuhl, Trevor Darrell
Photometric Stereo in a Scattering Medium [pdf]
Zak Murez, Tali Treibitz, Ravi Ramamoorthi, David Kriegman

Abstract: Photometric stereo is widely used for 3D reconstruction. However, its use in scattering media such as water, biological tissue and fog has been limited until now, because of forward scattered light from both the source and object, as well as light scattered back from the medium (backscatter). Here we make three contributions to address the key modes of light propagation, under the common single scattering assumption for dilute media. First, we show through extensive simulations that single-scattered light from a source can be approximated by a point light source with a single direction. This alleviates the need to handle light source blur explicitly. Next, we model the blur due to scattering of light from the object. We measure the object point-spread function and introduce a simple deconvolution method. Finally, we show how imaging fluorescence emission where available, eliminates the backscatter component and increases the signal-to-noise ratio. Experimental results in a water tank, with different concentrations of scattering media added, show that deconvolution produces higher-quality 3D reconstructions than previous techniques, and that when combined with fluorescence, can produce results similar to that in clear water even for highly turbid media.
Similar papers:
  • On Linear Structure From Motion for Light Field Cameras [pdf] - Ole Johannsen, Antonin Sulc, Bastian Goldluecke
  • Photogeometric Scene Flow for High-Detail Dynamic 3D Reconstruction [pdf] - Paulo F. U. Gotardo, Tomas Simon, Yaser Sheikh, Iain Matthews
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Nighttime Haze Removal With Glow and Multiple Light Colors [pdf] - Yu Li, Robby T. Tan, Michael S. Brown
  • Photometric Stereo With Small Angular Variations [pdf] - Jian Wang, Yasuyuki Matsushita, Boxin Shi, Aswin C. Sankaranarayanan
General Dynamic Scene Reconstruction From Multiple View Video [pdf]
Armin Mustafa, Hansung Kim, Jean-Yves Guillemaut, Adrian Hilton

Abstract: This paper introduces a general approach to dynamic scene reconstruction from multiple moving cameras without prior knowledge or limiting constraints on the scene structure, appearance, or illumination. Existing techniques for dynamic scene reconstruction from multiple wide-baseline camera views primarily focus on accurate reconstruction in controlled environments, where the cameras are fixed and calibrated and background is known. These approaches are not robust for general dynamic scenes captured with sparse moving cameras. Previous approaches for outdoor dynamic scene reconstruction assume prior knowledge of the static background appearance and structure. The primary contributions of this paper are twofold: an automatic method for initial coarse dynamic scene segmentation and reconstruction without prior knowledge of background appearance or structure; and a general robust approach for joint segmentation refinement and dense reconstruction of dynamic scenes from multiple wide-baseline static or moving cameras. Evaluation is performed on a variety of indoor and outdoor scenes with cluttered backgrounds and multiple dynamic non-rigid objects such as people. Comparison with state-of-the-art approaches demonstrates improved accuracy in both multiple view segmentation and dense reconstruction. The proposed approach also eliminates the requirement for prior knowledge of scene structure and appearance.
Similar papers:
  • Sparse Dynamic 3D Reconstruction From Unsynchronized Videos [pdf] - Enliang Zheng, Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
  • Joint Camera Clustering and Surface Segmentation for Large-Scale Multi-View Stereo [pdf] - Runze Zhang, Shiwei Li, Tian Fang, Siyu Zhu, Long Quan
  • Building Dynamic Cloud Maps From the Ground Up [pdf] - Calvin Murdock, Nathan Jacobs, Robert Pless
  • Exploiting Object Similarity in 3D Reconstruction [pdf] - Chen Zhou, Fatma Guney, Yizhou Wang, Andreas Geiger
Video Segmentation With Just a Few Strokes [pdf]
Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox

Abstract: As the use of videos is becoming more popular in computer vision, the need for annotated video datasets increases. Such datasets are required either as training data or simply as ground truth for benchmark datasets. A particular challenge in video segmentation is due to disocclusions, which hamper frame-to-frame propagation, in conjunction with non-moving objects. We show that a combination of motion from point trajectories, as known from motion segmentation, along with minimal supervision can largely help solve this problem. Moreover, we integrate a new constraint that enforces consistency of the color distribution in successive frames. We quantify user interaction effort with respect to segmentation quality on challenging ego motion videos. We compare our approach to a diverse set of algorithms in terms of user effort and in terms of performance on common video segmentation benchmarks.
Similar papers:
  • Fully Connected Object Proposals for Video Segmentation [pdf] - Federico Perazzi, Oliver Wang, Markus Gross, Alexander Sorkine-Hornung
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • Self-Occlusions and Disocclusions in Causal Video Object Segmentation [pdf] - Yanchao Yang, Ganesh Sundaramoorthi, Stefano Soatto
  • Motion Trajectory Segmentation via Minimum Cost Multicuts [pdf] - Margret Keuper, Bjoern Andres, Thomas Brox
Reflection Modeling for Passive Stereo [pdf]
Rahul Nair, Andrew Fitzgibbon, Daniel Kondermann, Carsten Rother

Abstract: Stereo reconstruction in presence of reality faces many challenges that still need to be addressed. This paper considers reflections, which introduce incorrect stereo matches due to the observation violating the diffuse-world assumption underlying the majority of stereo techniques. Unlike most existing work, which employ regularization or robust data terms to suppress such errors, we derive two least squares models from first principles that generalize diffuse world stereo and explicitly take reflections into account. These models are parametrized by depth, orientation and material properties, resulting in a total of up to 5 parameters per pixel that have to be estimated. Additionally widerange non-local interactions between viewed and reflected surface have to be taken into account. These two properties make optimization of the model appear prohibitive, but we present evidence that it is actually possible using a variant of patch match stereo.
Similar papers:
  • Photogeometric Scene Flow for High-Detail Dynamic 3D Reconstruction [pdf] - Paulo F. U. Gotardo, Tomas Simon, Yaser Sheikh, Iain Matthews
  • Variational PatchMatch MultiView Reconstruction and Refinement [pdf] - Philipp Heise, Brian Jensen, Sebastian Klose, Alois Knoll
  • Massively Parallel Multiview Stereopsis by Surface Normal Diffusion [pdf] - Silvano Galliani, Katrin Lasinger, Konrad Schindler
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Polarized 3D: High-Quality Depth Sensing With Polarization Cues [pdf] - Achuta Kadambi, Vage Taamazyan, Boxin Shi, Ramesh Raskar
Cutting Edge: Soft Correspondences in Multimodal Scene Parsing [pdf]
Sarah Taghavi Namin, Mohammad Najafi, Mathieu Salzmann, Lars Petersson

Abstract: Exploiting multiple modalities for semantic scene parsing has been shown to improve accuracy over the singlemodality scenario. Existing methods, however, assume that corresponding regions in two modalities have the same label. In this paper, we address the problem of data misalignment and label inconsistencies, e.g., due to moving objects, in semantic labeling, which violate the assumption of existing techniques. To this end, we formulate multimodal semantic labeling as inference in a CRF, and introduce latent nodes to explicitly model inconsistencies between two domains. These latent nodes allow us not only to leverage information from both domains to improve their labeling, but also to cut the edges between inconsistent regions. To eliminate the need for hand tuning the parameters of our model, we propose to learn intra-domain and inter-domain potential functions from training data. We demonstrate the benefits of our approach on two publicly available datasets containing 2D imagery and 3D point clouds. Thanks to our latent nodes and our learning strategy, our method outperforms the state-of-the-art in both cases.
Similar papers:
  • MANTRA: Minimum Maximum Latent Structural SVM for Image Classification and Ranking [pdf] - Thibaut Durand, Nicolas Thome, Matthieu Cord
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
  • Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf] - Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic
  • A Nonparametric Bayesian Approach Toward Stacked Convolutional Independent Component Analysis [pdf] - Sotirios P. Chatzis, Dimitrios Kosmopoulos
  • Similarity Gaussian Process Latent Variable Model for Multi-Modal Data Analysis [pdf] - Guoli Song, Shuhui Wang, Qingming Huang, Qi Tian
Direct Intrinsics: Learning Albedo-Shading Decomposition by Convolutional Regression [pdf]
Takuya Narihira, Michael Maire, Stella X. Yu

Abstract: We introduce a new approach to intrinsic image decomposition, the task of decomposing a single image into albedo and shading components. Our strategy, which we term direct intrinsics, is to learn a convolutional neural network (CNN) that directly predicts output albedo and shading channels from an input RGB image patch. Direct intrinsics is a departure from classical techniques for intrinsic image decomposition, which typically rely on physically-motivated priors and graph-based inference algorithms. The large-scale synthetic ground-truth of the MPI Sintel dataset plays a key role in training direct intrinsics. We demonstrate results on both the synthetic images of Sintel and the real images of the classic MIT intrinsic image dataset. On Sintel, direct intrinsics, using only RGB input, outperforms all prior work, including methods that rely on RGB+Depth input. Direct intrinsics also generalizes across modalities; it produces quite reasonable decompositions on the real images of the MIT dataset. Our results indicate that the marriage of CNNs with synthetic training data may be a powerful new technique for tackling classic problems in computer vision.
Similar papers:
  • Learning Ordinal Relationships for Mid-Level Vision [pdf] - Daniel Zoran, Phillip Isola, Dilip Krishnan, William T. Freeman
  • A Comprehensive Multi-Illuminant Dataset for Benchmarking of the Intrinsic Image Algorithms [pdf] - Shida Beigpour, Andreas Kolb, Sven Kunz
  • Intrinsic Decomposition of Image Sequences From Local Temporal Variations [pdf] - Pierre-Yves Laffont, Jean-Charles Bazin
  • Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf] - Naejin Kong, Michael J. Black
  • Intrinsic Scene Decomposition From RGB-D images [pdf] - Mohammed Hachama, Bernard Ghanem, Peter Wonka
Bayesian Non-Parametric Inference for Manifold Based MoCap Representation [pdf]
Fabrizio Natola, Valsamis Ntouskos, Marta Sanzari, Fiora Pirri

Abstract: We propose a novel approach to human action recognition, with motion capture data (MoCap), based on grouping sub-body parts. By representing configurations of actions as manifolds, joint positions are mapped on a subspace via principal geodesic analysis. The reduced space is still highly informative and allows for classification based on a non-parametric Bayesian approach, generating behaviors for each sub-body part. Having partitioned the set of joints, poses relative to a sub-body part are exchangeable, given a specified prior and can elicit, in principle, infinite behaviors. The generation of these behaviors is specified by a Dirichlet process mixture. We show with several experiments that the recognition gives very promising results, outperforming methods requiring temporal alignment.
Similar papers:
  • Action Localization in Videos Through Context Walk [pdf] - Khurram Soomro, Haroon Idrees, Mubarak Shah
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
Dense Image Registration and Deformable Surface Reconstruction in Presence of Occlusions and Minimal Texture [pdf]
Dat Tien Ngo, Sanghyuk Park, Anne Jorstad, Alberto Crivellaro, Chang D. Yoo, Pascal Fua

Abstract: Deformable surface tracking from monocular images is well-known to be under-constrained. Occlusions often make the task even more challenging, and can result in failure if the surface is not sufficiently textured. In this work, we explicitly address the problem of 3D reconstruction of poorly textured, occluded surfaces, proposing a framework based on a template-matching approach that scales dense robust features by a relevancy score. Our approach is extensively compared to current methods employing both local feature matching and dense template alignment. We test on standard datasets as well as on a new dataset (that will be made publicly available) of a sparsely textured, occluded surface. Our framework achieves state-of-the-art results for both well and poorly textured, occluded surfaces.
Similar papers:
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Global, Dense Multiscale Reconstruction for a Billion Points [pdf] - Benjamin Ummenhofer, Thomas Brox
  • Exploiting Object Similarity in 3D Reconstruction [pdf] - Chen Zhou, Fatma Guney, Yizhou Wang, Andreas Geiger
  • Peeking Template Matching for Depth Extension [pdf] - Simon Korman, Eyal Ofek, Shai Avidan
  • Direct, Dense, and Deformable: Template-Based Non-Rigid 3D Reconstruction From RGB Video [pdf] - Rui Yu, Chris Russell, Neill D. F. Campbell, Lourdes Agapito
An MRF-Poselets Model for Detecting Highly Articulated Humans [pdf]
Duc Thanh Nguyen, Minh-Khoi Tran, Sai-Kit Yeung

Abstract: Detecting highly articulated objects such as humans is a challenging problem. This paper proposes a novel partbased model built upon poselets, a notion of parts, and Markov Random Field (MRF) for modelling the human body structure under the variation of human poses and viewpoints. The problem of human detection is then formulated as maximum a posteriori (MAP) estimation in the MRF model. Variational mean field method, a robust statistical inference, is adopted to approximate the MAP estimation. The proposed method was evaluated and compared with existing methods on different test sets including H3D and PASCAL VOC 2007-2009. Experimental results have favourbly shown the robustness of the proposed method in comparison to the state-of-the-art. 2 minhkhoi tran@mymail.sutd.edu.sg which the human object image (detection window) is represented in a regular grid of blocks. Features, e.g. histograms of oriented gradients (HOG) [11], are then extracted from the blocks and concatenated to form a feature vector describing the whole human object. The success of this approach has been confirmed in detecting human objects with less variation of poses and viewpoints, e.g. pedestrians. On the other hand, local methods model the human object as a set of parts, e.g. [25, 7, 1, 15, 37, 38, 19, 8, 20]. Compared with the global approach, the local approach is more robust to model the high articulation of the human body. However, the performance of local methods significantly depends on detectin
Similar papers:
  • A Nonparametric Bayesian Approach Toward Stacked Convolutional Independent Component Analysis [pdf] - Sotirios P. Chatzis, Dimitrios Kosmopoulos
  • Deep Learning Strong Parts for Pedestrian Detection [pdf] - Yonglong Tian, Ping Luo, Xiaogang Wang, Xiaoou Tang
  • Beyond Tree Structure Models: A New Occlusion Aware Graphical Model for Human Pose Estimation [pdf] - Lianrui Fu, Junge Zhang, Kaiqi Huang
  • Human Pose Estimation in Videos [pdf] - Dong Zhang, Mubarak Shah
  • Actions and Attributes From Wholes and Parts [pdf] - Georgia Gkioxari, Ross Girshick, Jitendra Malik
Fast and Effective L0 Gradient Minimization by Region Fusion [pdf]
Rang M. H. Nguyen, Michael S. Brown

Abstract: L0 gradient minimization can be applied to an input signal to control the number of non-zero gradients. This is useful in reducing small gradients generally associated with signal noise, while preserving important signal features. In computer vision, L0 gradient minimization has found applications in image denoising, 3D mesh denoising, and image enhancement. Minimizing the L0 norm, however, is an NP-hard problem because of its non-convex property. As a result, existing methods rely on approximation strategies to perform the minimization. In this paper, we present a new method to perform L0 gradient minimization that is fast and effective. Our method uses a descent approach based on region fusion that converges faster than other methods while providing a better approximation of the optimal L0 norm. In addition, our method can be applied to both 2D images and 3D mesh topologies. The effectiveness of our approach is demonstrated on a number of examples.
Similar papers:
  • Pan-Sharpening With a Hyper-Laplacian Penalty [pdf] - Yiyong Jiang, Xinghao Ding, Delu Zeng, Yue Huang, John Paisley
  • Mutual-Structure for Joint Filtering [pdf] - Xiaoyong Shen, Chao Zhou, Li Xu, Jiaya Jia
  • A Wavefront Marching Method for Solving the Eikonal Equation on Cartesian Grids [pdf] - Brais Cancela, Marcos Ortega, Manuel G. Penedo
  • Learning Nonlinear Spectral Filters for Color Image Reconstruction [pdf] - Michael Moeller, Julia Diebold, Guy Gilboa, Daniel Cremers
  • POP Image Fusion - Derivative Domain Image Fusion Without Reintegration [pdf] - Graham D. Finlayson, Alex E. Hayes
The One Triangle Three Parallelograms Sampling Strategy and Its Application in Shape Regression [pdf]
Mikael Nilsson

Abstract: The purpose of this paper is threefold. Firstly, the paper introduces the One Triangle Three Parallelograms (OTTP) sampling strategy, which can be viewed as a way to index pixels from a given shape and image. Secondly, a framework for cascaded shape regression, including the OTTP sampling, is presented. In short, this framework involves binary pixel tests for appearance features combined with shape features followed by a large linear system for each regression stage in the cascade. The proposed solution is found to produce state-of-the-art results on the task of facial landmark estimation. Thirdly, the dependence of accuracy of the landmark predictions and the placement of the mean shape within the detection box is discussed and a method to visualize it is presented.
Similar papers:
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • Leveraging Datasets With Varying Annotations for Face Alignment via Deep Regression Network [pdf] - Jie Zhang, Meina Kan, Shiguang Shan, Xilin Chen
  • Compositional Hierarchical Representation of Shape Manifolds for Classification of Non-Manifold Shapes [pdf] - Mete Ozay, Umit Rusen Aktas, Jeremy L. Wyatt, Ales Leonardis
  • Single Image Pop-Up From Discriminatively Learned Parts [pdf] - Menglong Zhu, Xiaowei Zhou, Kostas Daniilidis
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
A Linear Generalized Camera Calibration From Three Intersecting Reference Planes [pdf]
Mai Nishimura, Shohei Nobuhara, Takashi Matsuyama, Shinya Shimizu, Kensaku Fujii

Abstract: Water Tank Camera 4 This paper presents a new generalized (or ray-pixel, raxel) camera calibration algorithm for camera systems involving distortions by unknown refraction and reflection processes. The key idea is use of intersections of calibration planes, while conventional methods utilized collinearity constraints of points on the planes. We show that intersections of calibration planes can realize a simple linear algorithm, and that our method can be applied to any ray-distributions while conventional methods require knowing the ray-distribution class in advance. Evaluations using synthesized and real datasets demonstrate the performance of our method quantitatively and qualitatively.
Similar papers:
  • Structure From Motion Using Structure-Less Resection [pdf] - Enliang Zheng, Changchang Wu
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
  • On the Equivalence of Moving Entrance Pupil and Radial Distortion for Camera Calibration [pdf] - Avinash Kumar, Narendra Ahuja
  • Non-Parametric Structure-Based Calibration of Radially Symmetric Cameras [pdf] - Federico Camposeco, Torsten Sattler, Marc Pollefeys
  • Accurate Camera Calibration Robust to Defocus Using a Smartphone [pdf] - Hyowon Ha, Yunsu Bok, Kyungdon Joo, Jiyoung Jung, In So Kweon
Multi-View Domain Generalization for Visual Recognition [pdf]
Li Niu, Wen Li, Dong Xu

Abstract: In this paper, we propose a new multi-view domain generalization (MVDG) approach for visual recognition, in which we aim to use the source domain samples with multiple types of features (i.e., multi-view features) to learn robust classifiers that can generalize well to any unseen target domain. Considering the recent works show the domain generalization capability can be enhanced by fusing multiple SVM classifiers, we build upon exemplar SVMs to learn a set of SVM classifiers by using one positive sample and all negative samples in the source domain each time. When the source domain samples come from multiple latent domains, we expect the weight vectors of exemplar SVM classifiers can be organized into multiple hidden clusters. To exploit such cluster structure, we organize the weight vectors learnt on each view as a weight matrix and seek the low-rank representation by reconstructing this weight matrix using itself as the dictionary. To enforce the consistency of inherent cluster structures discovered from the weight matrices learnt on different views, we introduce a new regularizer to minimize the mismatch between any two representation matrices on different views. We also develop an efficient alternating optimization algorithm and further extend our MVDG approach for domain adaptation by exploiting the manifold structure of unlabeled target domain samples. Comprehensive experiments for visual recognition clearly demonstrate the effectiveness of our approaches for domain ge
Similar papers:
  • Unsupervised Domain Adaptation for Zero-Shot Learning [pdf] - Elyor Kodirov, Tao Xiang, Zhenyong Fu, Shaogang Gong
  • Domain Generalization for Object Recognition With Multi-Task Autoencoders [pdf] - Muhammad Ghifary, W. Bastiaan Kleijn, Mengjie Zhang, David Balduzzi
  • Simultaneous Deep Transfer Across Domains and Tasks [pdf] - Eric Tzeng, Judy Hoffman, Trevor Darrell, Kate Saenko
  • Unsupervised Domain Adaptation With Imbalanced Cross-Domain Data [pdf] - Tzu Ming Harry Hsu, Wei Yu Chen, Cheng-An Hou, Yao-Hung Hubert Tsai, Yi-Ren Yeh, Yu-Chiang Frank Wang
  • Bi-Shifting Auto-Encoder for Unsupervised Domain Adaptation [pdf] - Meina Kan, Shiguang Shan, Xilin Chen
Learning Deconvolution Network for Semantic Segmentation [pdf]
Hyeonwoo Noh, Seunghoon Hong, Bohyung Han

Abstract: car car car person bicycle We propose a novel semantic segmentation algorithm by learning a deep deconvolution network. We learn the network on top of the convolutional layers adopted from VGG 16-layer net. The deconvolution network is composed of deconvolution and unpooling layers, which identify pixelwise class labels and predict segmentation masks. We apply the trained network to each proposal in an input image, and construct the final semantic segmentation map by combining the results from all proposals in a simple manner. The proposed algorithm mitigates the limitations of the existing methods based on fully convolutional networks by integrating deep deconvolution network and proposal-wise prediction; our segmentation method typically identifies detailed structures and handles objects in multiple scales naturally. Our network demonstrates outstanding performance in PASCAL VOC 2012 dataset, and we achieve the best accuracy (72.5%) among the methods trained without using Microsoft COCO dataset through ensemble with the fully convolutional network.
Similar papers:
  • Predicting Depth, Surface Normals and Semantic Labels With a Common Multi-Scale Convolutional Architecture [pdf] - David Eigen, Rob Fergus
  • BoxSup: Exploiting Bounding Boxes to Supervise Convolutional Networks for Semantic Segmentation [pdf] - Jifeng Dai, Kaiming He, Jian Sun
  • Human Parsing With Contextualized Convolutional Neural Network [pdf] - Xiaodan Liang, Chunyan Xu, Xiaohui Shen, Jianchao Yang, Si Liu, Jinhui Tang, Liang Lin, Shuicheng Yan
  • Joint Object and Part Segmentation Using Deep Learned Potentials [pdf] - Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille
  • Conditional Random Fields as Recurrent Neural Networks [pdf] - Shuai Zheng, Sadeep Jayasumana, Bernardino Romera-Paredes, Vibhav Vineet, Zhizhong Su, Dalong Du, Chang Huang, Philip H. S. Torr
Cascaded Sparse Spatial Bins for Efficient and Effective Generic Object Detection [pdf]
David Novotny, Jiri Matas

Abstract: A novel efficient method for extraction of object proposals is introduced. Its objectness function exploits deep spatial pyramid features, a novel fast-to-compute HoGbased edge statistic and the EdgeBoxes score [42]. The efficiency is achieved by the use of spatial bins in a novel combination with sparsity-inducing group normalized SVM. State-of-the-art recall performance is achieved on Pascal VOC07, significantly outperforming methods with comparable speed. Interestingly, when only 100 proposals per image are considered the method attains 78% recall on VOC07. The method improves mAP of the RCNN stateof-the-art class-specific detector, increasing it by 10 points when only 50 proposals are used in each image. The system trained on twenty classes performs well on the two hundred class ILSVRC2013 set confirming generalization capability.
Similar papers:
  • Contour Box: Rejecting Object Proposals Without Explicit Closed Contours [pdf] - Cewu Lu, Shu Liu, Jiaya Jia, Chi-Keung Tang
  • Box Aggregation for Proposal Decimation: Last Mile of Object Detection [pdf] - Shu Liu, Cewu Lu, Jiaya Jia
  • Oriented Object Proposals [pdf] - Shengfeng He, Rynson W.H. Lau
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • DeepProposal: Hunting Objects by Cascading Deep Convolutional Layers [pdf] - Amir Ghodrati, Ali Diba, Marco Pedersoli, Tinne Tuytelaars, Luc Van Gool
Component-Wise Modeling of Articulated Objects [pdf]
Valsamis Ntouskos, Marta Sanzari, Bruno Cafaro, Federico Nardi, Fabrizio Natola, Fiora Pirri, Manuel Ruiz

Abstract: We introduce a novel framework for modeling articulated objects based on the aspects of their components. By decomposing the object into components, we divide the problem in smaller modeling tasks. After obtaining 3D models for each component aspect by employing a shape deformation paradigm, we merge them together, forming the object components. The final model is obtained by assembling the components using an optimization scheme which fits the respective 3D models to the corresponding apparent contours in a reference pose. The results suggest that our approach can produce realistic 3D models of articulated objects in reasonable time.
Similar papers:
  • Contour Flow: Middle-Level Motion Estimation by Combining Motion Segmentation and Contour Alignment [pdf] - Huijun Di, Qingxuan Shi, Feng Lv, Ming Qin, Yao Lu
  • Semantic Component Analysis [pdf] - Calvin Murdock, Fernando De la Torre
  • Resolving Scale Ambiguity Via XSlit Aspect Ratio Analysis [pdf] - Wei Yang, Haiting Lin, Sing Bing Kang, Jingyi Yu
  • Interactive Visual Hull Refinement for Specular and Transparent Object Surface Reconstruction [pdf] - Xinxin Zuo, Chao Du, Sen Wang, Jiangbin Zheng, Ruigang Yang
  • Thin Structure Estimation With Curvature Regularization [pdf] - Dmitrii Marin, Yuchen Zhong, Maria Drangova, Yuri Boykov
Towards Pointless Structure From Motion: 3D Reconstruction and Camera Parameters From General 3D Curves [pdf]
Irina Nurutdinova, Andrew Fitzgibbon

Abstract: Modern structure from motion (SfM) remains dependent on point features to recover camera positions, meaning that reconstruction is severely hampered in low-texture environments, for example scanning a plain coffee cup on an uncluttered table. We show how 3D curves can be used to refine camera position estimation in challenging low-texture scenes. In contrast to previous work, we allow the curves to be partially observed in all images, meaning that for the first time, curve-based SfM can be demonstrated in realistic scenes. The algorithm is based on bundle adjustment, so needs an initial estimate, but even a poor estimate from a few point correspondences can be substantially improved by including curves, suggesting that this method would benefit many existing systems.
Similar papers:
  • Joint Camera Clustering and Surface Segmentation for Large-Scale Multi-View Stereo [pdf] - Runze Zhang, Shiwei Li, Tian Fang, Siyu Zhu, Long Quan
  • Accurate Camera Calibration Robust to Defocus Using a Smartphone [pdf] - Hyowon Ha, Yunsu Bok, Kyungdon Joo, Jiyoung Jung, In So Kweon
  • A Linear Generalized Camera Calibration From Three Intersecting Reference Planes [pdf] - Mai Nishimura, Shohei Nobuhara, Takashi Matsuyama, Shinya Shimizu, Kensaku Fujii
  • General Dynamic Scene Reconstruction From Multiple View Video [pdf] - Armin Mustafa, Hansung Kim, Jean-Yves Guillemaut, Adrian Hilton
  • High Quality Structure From Small Motion for Rolling Shutter Cameras [pdf] - Sunghoon Im, Hyowon Ha, Gyeongmin Choe, Hae-Gon Jeon, Kyungdon Joo, In So Kweon
Training a Feedback Loop for Hand Pose Estimation [pdf]
Markus Oberweger, Paul Wohlhart, Vincent Lepetit

Abstract: Add update We propose an entirely data-driven approach to estimating the 3D pose of a hand given a depth image. We show that we can correct the mistakes made by a Convolutional Neural Network trained to predict an estimate of the 3D pose by using a feedback loop. The components of this feedback loop are also Deep Networks, optimized using training data. They remove the need for fitting a 3D model to the input data, which requires both a carefully designed fitting function and algorithm. We show that our approach outperforms state-of-the-art methods, and is efficient as our implementation runs at over 400 fps on a single GPU. Pose 2 3 Synthesizer CNN Initialize Synthesized Image Updater CNN 1 Predictor CNN Input Image
Similar papers:
  • 3D Hand Pose Estimation Using Randomized Decision Forest With Segmentation Index Points [pdf] - Peiyi Li, Haibin Ling, Xi Li, Chunyuan Liao
  • Understanding Everyday Hands in Action From RGB-D Images [pdf] - Gregory Rogez, James S. Supancic III, Deva Ramanan
  • Opening the Black Box: Hierarchical Sampling Optimization for Estimating Human Hand Pose [pdf] - Danhang Tang, Jonathan Taylor, Pushmeet Kohli, Cem Keskin, Tae-Kyun Kim, Jamie Shotton
  • Depth-Based Hand Pose Estimation: Data, Methods, and Challenges [pdf] - James S. Supancic III, Gregory Rogez, Yi Yang, Jamie Shotton, Deva Ramanan
  • A Collaborative Filtering Approach to Real-Time Hand Pose Estimation [pdf] - Chiho Choi, Ayan Sinha, Joon Hee Choi, Sujin Jang, Karthik Ramani
Person Recognition in Personal Photo Collections [pdf]
Seong Joon Oh, Rodrigo Benenson, Mario Fritz, Bernt Schiele

Abstract: Recognising persons in everyday photos presents major challenges (occluded faces, different clothing, locations, etc.) for machine vision. We propose a convnet based person recognition system on which we provide an in-depth analysis of informativeness of different body cues, impact of training data, and the common failure modes of the system. In addition, we discuss the limitations of existing benchmarks and propose more challenging ones. Our method is simple and is built on open source and open data, yet it improves the state of the art results on a large dataset of social media photos (PIPA).
Similar papers:
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Deep Learning Face Attributes in the Wild [pdf] - Ziwei Liu, Ping Luo, Xiaogang Wang, Xiaoou Tang
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
  • Uncovering Interactions and Interactors: Joint Estimation of Head, Body Orientation and F-Formations From Surveillance Videos [pdf] - Elisa Ricci, Jagannadan Varadarajan, Ramanathan Subramanian, Samuel Rota Bulo, Narendra Ahuja, Oswald Lanz
  • Robust Model-Based 3D Head Pose Estimation [pdf] - Gregory P. Meyer, Shalini Gupta, Iuri Frosio, Dikpal Reddy, Jan Kautz
Learning Deep Representation With Large-Scale Attributes [pdf]
Wanli Ouyang, Hongyang Li, Xingyu Zeng, Xiaogang Wang

Abstract: Learning strong feature representations from large scale supervision has achieved remarkable success in computer vision as the emergence of deep learning techniques. It is driven by big visual data with rich annotations. This paper contributes a large-scale object attribute database 1 that contains rich attribute annotations (over 300 attributes) for 180k samples and 494 object classes. Based on the ImageNet object detection dataset, it annotates the rotation, viewpoint, object part location, part occlusion, part existence, common attributes, and class-specific attributes. Then we use this dataset to train deep representations and extensively evaluate how these attributes are useful on the general object detection task. In order to make better use of the attribute annotations, a deep learning scheme is proposed by modeling the relationship of attributes and hierarchically clustering them into semantically meaningful mixture types. Experimental results show that the attributes are helpful in learning better features and improving the object detection accuracy by 2.6% in mAP on the ILSVRC 2014 object detection dataset and 2.4% in mAP on PASCAL VOC 2007 object detection dataset. Such improvement is well generalized across datasets.
Similar papers:
  • Just Noticeable Differences in Visual Attributes [pdf] - Aron Yu, Kristen Grauman
  • Multi-Task Learning With Low Rank Attribute Embedding for Person Re-Identification [pdf] - Chi Su, Fan Yang, Shiliang Zhang, Qi Tian, Larry S. Davis, Wen Gao
  • Discovering the Spatial Extent of Relative Attributes [pdf] - Fanyi Xiao, Yong Jae Lee
  • Attributed Grammars for Joint Estimation of Human Attributes, Part and Pose [pdf] - Seyoung Park, Song-Chun Zhu
  • A Unified Multiplicative Framework for Attribute Learning [pdf] - Kongming Liang, Hong Chang, Shiguang Shan, Xilin Chen
Compositional Hierarchical Representation of Shape Manifolds for Classification of Non-Manifold Shapes [pdf]
Mete Ozay, Umit Rusen Aktas, Jeremy L. Wyatt, Ales Leonardis

Abstract: We address the problem of statistical learning of shape models which are invariant to translation, rotation and scale in compositional hierarchies when data spaces of measurements and shape spaces are not topological manifolds. In practice, this problem is observed while modeling shapes having multiple disconnected components, e.g. partially occluded shapes in cluttered scenes. We resolve the aforementioned problem by first reformulating the relationship between data and shape spaces considering the interaction between Receptive Fields (RFs) and Shape Manifolds (SMs) in a compositional hierarchical shape vocabulary. Then, we suggest a method to model the topological structure of the SMs for statistical learning of the geometric transformations of the shapes that are defined by group actions on the SMs. For this purpose, we design a disjoint union topology using an indexing mechanism for the formation of shape models on SMs in the vocabulary, recursively. We represent the topological relationship between shape components using graphs, which are aggregated to construct a hierarchical graph structure for the shape vocabulary. To this end, we introduce a framework to implement the indexing mechanisms for the employment of the vocabulary for structural shape classification. The proposed approach is used to construct invariant shape representations. Results on benchmark shape classification outperform state-of-the-art methods.
Similar papers:
  • Detailed Full-Body Reconstructions of Moving People From Monocular RGB-D Sequences [pdf] - Federica Bogo, Michael J. Black, Matthew Loper, Javier Romero
  • BodyPrint: Pose Invariant 3D Shape Matching of Human Bodies [pdf] - Jiangping Wang, Kai Ma, Vivek Kumar Singh, Thomas Huang, Terrence Chen
  • Multi-View Convolutional Neural Networks for 3D Shape Recognition [pdf] - Hang Su, Subhransu Maji, Evangelos Kalogerakis, Erik Learned-Miller
  • Learning Shape, Motion and Elastic Models in Force Space [pdf] - Antonio Agudo, Francesc Moreno-Noguer
  • The One Triangle Three Parallelograms Sampling Strategy and Its Application in Shape Regression [pdf] - Mikael Nilsson
Weakly- and Semi-Supervised Learning of a Deep Convolutional Network for Semantic Image Segmentation [pdf]
George Papandreou, Liang-Chieh Chen, Kevin P. Murphy, Alan L. Yuille

Abstract: Deep convolutional neural networks (DCNNs) trained on a large number of images with strong pixel-level annotations have recently significantly pushed the state-of-art in semantic image segmentation. We study the more challenging problem of learning DCNNs for semantic image segmentation from either (1) weakly annotated training data such as bounding boxes or image-level labels or (2) a combination of few strongly labeled and many weakly labeled images, sourced from one or multiple datasets. We develop Expectation-Maximization (EM) methods for semantic image segmentation model training under these weakly supervised and semi-supervised settings. Extensive experimental evaluation shows that the proposed techniques can learn models delivering competitive results on the challenging PASCAL VOC 2012 image segmentation benchmark, while requiring significantly less annotation effort. We share source code implementing the proposed system at https: //bitbucket.org/deeplab/deeplab-public.
Similar papers:
  • AttentionNet: Aggregating Weak Directions for Accurate Object Detection [pdf] - Donggeun Yoo, Sunggyun Park, Joon-Young Lee, Anthony S. Paek, In So Kweon
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • Augmenting Strong Supervision Using Web Data for Fine-Grained Categorization [pdf] - Zhe Xu, Shaoli Huang, Ya Zhang, Dacheng Tao
  • Constrained Convolutional Neural Networks for Weakly Supervised Segmentation [pdf] - Deepak Pathak, Philipp Krahenbuhl, Trevor Darrell
  • BoxSup: Exploiting Bounding Boxes to Supervise Convolutional Networks for Semantic Segmentation [pdf] - Jifeng Dai, Kaiming He, Jian Sun
Semantic Pose Using Deep Networks Trained on Synthetic RGB-D [pdf]
Jeremie Papon, Markus Schoeler

Abstract: Synthetic Input 2x2 Max Pool Conv. 3x3 2x2 Max Pool Conv. 1x1 Conv. 1x1 Conv. 3x3 Conv. 1x1 Conv. 5x5 DepthConcat 2x2 Max Pool Conv. 1x1 Conv. 3x3 Conv. 1x1 Conv. 5x5 Object Proposals Conv. 1x1 DepthConcat FC Layer FC Layer FC Layer FC Layer SoftMax Concat FC Layer Position Output Concat Estimated Semantic Pose FC Layer In this work we address the problem of indoor scene understanding from RGB-D images. Specifically, we propose to find instances of common furniture classes, their spatial extent, and their pose with respect to generalized class models. To accomplish this, we use a deep, wide, multi-output convolutional neural network (CNN) that predicts class, pose, and location of possible objects simultaneously. To overcome the lack of large annotated RGB-D training sets (especially those with pose), we use an on-the-fly rendering pipeline that generates realistic cluttered room scenes in parallel to training. We then perform transfer learning on the relatively small amount of publicly available annotated RGB-D data, and find that our model is able to successfully annotate even highly challenging real scenes. Importantly, our trained network is able to understand noisy and sparse observations of highly cluttered scenes with a remarkable degree of accuracy, inferring class and pose from a very limited set of cues. Additionally, our neural network is only moderately deep and computes class, pose and position in tandem, so the overall run-time is significantly
Similar papers:
  • Learning Deep Object Detectors From 3D Models [pdf] - Xingchao Peng, Baochen Sun, Karim Ali, Kate Saenko
  • Learning Analysis-by-Synthesis for 6D Pose Estimation in RGB-D Images [pdf] - Alexander Krull, Eric Brachmann, Frank Michel, Michael Ying Yang, Stefan Gumhold, Carsten Rother
  • Training a Feedback Loop for Hand Pose Estimation [pdf] - Markus Oberweger, Paul Wohlhart, Vincent Lepetit
  • Pose Induction for Novel Object Categories [pdf] - Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
Efficient Classifier Training to Minimize False Merges in Electron Microscopy Segmentation [pdf]
Toufiq Parag, Dan C. Ciresan, Alessandro Giusti

Abstract: The prospect of neural reconstruction from Electron Microscopy (EM) images has been elucidated by the automatic segmentation algorithms. Although segmentation algorithms eliminate the necessity of tracing the neurons by hand, significant manual effort is still essential for correcting the mistakes they make. A considerable amount of human labor is also required for annotating groundtruth volumes for training the classifiers of a segmentation framework. It is critically important to diminish the dependence on human interaction in the overall reconstruction system. This study proposes a novel classifier training algorithm for EM segmentation aimed to reduce the amount of manual effort demanded by the groundtruth annotation and error refinement tasks. Instead of using an exhaustive pixel level groundtruth, an active learning algorithm is proposed for sparse labeling of pixel and boundaries of superpixels. Because over-segmentation errors are in general more tolerable and easier to correct than the under-segmentation errors, our algorithm is designed to prioritize minimization of false-merges over false-split mistakes. Our experiments on both 2D and 3D data suggest that the proposed method yields segmentation outputs that are more amenable to neural reconstruction than those of existing methods.
Similar papers:
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • Introducing Geometry in Active Learning for Image Segmentation [pdf] - Ksenia Konyushkova, Raphael Sznitman, Pascal Fua
  • High-for-Low and Low-for-High: Efficient Boundary Detection From Deep Object Features and its Applications to High-Level Vision [pdf] - Gedas Bertasius, Jianbo Shi, Lorenzo Torresani
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
  • Efficient Video Segmentation Using Parametric Graph Partitioning [pdf] - Chen-Ping Yu, Hieu Le, Gregory Zelinsky, Dimitris Samaras
As-Rigid-As-Possible Volumetric Shape-From-Template [pdf]
Shaifali Parashar, Daniel Pizarro, Adrien Bartoli, Toby Collins

Abstract: The objective of Shape-from-Template (SfT) is to infer an objects shape from a single image and a 3D object template. Existing methods are called thin-shell SfT as they represent the object by its outer surface. This may be an open surface for thin objects such as a piece of paper or a closed surface for thicker objects such as a ball. We propose volumetric SfT, which specifically handles objects of the latter kind. Volumetric SfT uses the objects full volume to express the deformation constraints and reconstructs the objects surface and interior deformation. This is a challenging problem because for opaque objects, only a part of the outer surface is visible in the image. Inspired by meshediting techniques, we use an As-Rigid-As-Possible (ARAP) deformation model that softly imposes local rigidity. We formalise ARAP isometric SfT as a constrained variational optimisation problem which we solve using iterative optimisation. We present strategies to find an initial solution based on thin-shell SfT and volume propagation. Experiments with synthetic and real data show that our method has a typical maximum relative error of 5% in reconstructing the deformation of an entire object, including its back and interior for which no visual data is available.
Similar papers:
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Dense Image Registration and Deformable Surface Reconstruction in Presence of Occlusions and Minimal Texture [pdf] - Dat Tien Ngo, Sanghyuk Park, Anne Jorstad, Alberto Crivellaro, Chang D. Yoo, Pascal Fua
  • Direct, Dense, and Deformable: Template-Based Non-Rigid 3D Reconstruction From RGB Video [pdf] - Rui Yu, Chris Russell, Neill D. F. Campbell, Lourdes Agapito
  • Peeking Template Matching for Depth Extension [pdf] - Simon Korman, Eyal Ofek, Shai Avidan
  • Exploiting Object Similarity in 3D Reconstruction [pdf] - Chen Zhou, Fatma Guney, Yizhou Wang, Andreas Geiger
Attributed Grammars for Joint Estimation of Human Attributes, Part and Pose [pdf]
Seyoung Park, Song-Chun Zhu

Abstract: In this paper, we are interested in developing compositional models to explicit representing pose, parts and attributes and tackling the tasks of attribute recognition, pose estimation and part localization jointly. This is different from the recent trend of using CNN-based approaches for training and testing on these tasks separately with a large amount of data. Conventional attribute models typically use a large number of region-based attribute classifiers on parts of pre-trained pose estimator without explicitly detecting the object or its parts, or considering the correlations between attributes. In contrast, our approach jointly represents both the object parts and their semantic attributes within a unified compositional hierarchy. We apply our attributed grammar model to the task of human parsing by simultaneously performing part localization and attribute recognition. We show our modeling helps performance improvements on pose-estimation task and also outperforms on other existing methods on attribute prediction task.
Similar papers:
  • Multi-Task Learning With Low Rank Attribute Embedding for Person Re-Identification [pdf] - Chi Su, Fan Yang, Shiliang Zhang, Qi Tian, Larry S. Davis, Wen Gao
  • Just Noticeable Differences in Visual Attributes [pdf] - Aron Yu, Kristen Grauman
  • Discovering the Spatial Extent of Relative Attributes [pdf] - Fanyi Xiao, Yong Jae Lee
  • Learning Deep Representation With Large-Scale Attributes [pdf] - Wanli Ouyang, Hongyang Li, Xingyu Zeng, Xiaogang Wang
  • A Unified Multiplicative Framework for Attribute Learning [pdf] - Kongming Liang, Hong Chang, Shiguang Shan, Xilin Chen
Constrained Convolutional Neural Networks for Weakly Supervised Segmentation [pdf]
Deepak Pathak, Philipp Krahenbuhl, Trevor Darrell

Abstract: We present an approach to learn a dense pixel-wise labeling from image-level tags. Each image-level tag imposes constraints on the output labeling of a Convolutional Neural Network (CNN) classifier. We propose Constrained CNN (CCNN), a method which uses a novel loss function to optimize for any set of linear constraints on the output space (i.e. predicted label distribution) of a CNN. Our loss formulation is easy to optimize and can be incorporated directly into standard stochastic gradient descent optimization. The key idea is to phrase the training objective as a biconvex optimization for linear models, which we then relax to nonlinear deep networks. Extensive experiments demonstrate the generality of our new learning framework. The constrained loss yields state-of-the-art results on weakly supervised semantic image segmentation. We further demonstrate that adding slightly more supervision can greatly improve the performance of the learning algorithm. Input Image Predicted labeling Person CNN Car Network Output Person Car Weak Labels Latent Distribution Constrained Region Figure 1: We train convolutional neural networks from a set of linear constraints on the output variables. The network output is encouraged to follow a latent probability distribution, which lies in the constraint manifold. The resulting loss is easy to optimize and can incorporate arbitrary linear constraints.
Similar papers:
  • Semantic Segmentation of RGBD Images With Mutex Constraints [pdf] - Zhuo Deng, Sinisa Todorovic, Longin Jan Latecki
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • Semi-Supervised Normalized Cuts for Image Segmentation [pdf] - Selene E. Chew, Nathan D. Cahill
  • Semantic Component Analysis [pdf] - Calvin Murdock, Fernando De la Torre
  • Weakly- and Semi-Supervised Learning of a Deep Convolutional Network for Semantic Image Segmentation [pdf] - George Papandreou, Liang-Chieh Chen, Kevin P. Murphy, Alan L. Yuille
Robust and Optimal Sum-of-Squares-Based Point-to-Plane Registration of Image Sets and Structured Scenes [pdf]
Danda Pani Paudel, Adlane Habed, Cedric Demonceaux, Pascal Vasseur

Abstract: This paper deals with the problem of registering a known structured 3D scene and its metric Structure-from-Motion (SfM) counterpart. The proposed work relies on a prior plane segmentation of the 3D scene and aligns the data obtained from both modalities by solving the point-toplane assignment problem. An inliers-maximization approach within a Branch-and-Bound (BnB) search scheme is adopted. For the first time in this paper, a Sum-of-Squares optimization theory framework is employed for identifying point-to-plane mismatches (i.e. outliers) with certainty. This allows us to iteratively build potential inliers sets and converge to the solution satisfied by the largest number of point-to-plane assignments. Furthermore, our approach is boosted by new plane visibility conditions which are also introduced in this paper. Using this framework, we solve the registration problem in two cases: (i) a set of putative point-to-plane correspondences (with possibly overwhelmingly many outliers) is given as input and (ii) no initial correspondences are given. In both cases, our approach yields outstanding results in terms of robustness and optimality.
Similar papers:
  • Structure From Motion Using Structure-Less Resection [pdf] - Enliang Zheng, Changchang Wu
  • The Joint Image Handbook [pdf] - Matthew Trager, Martial Hebert, Jean Ponce
  • Registering Images to Untextured Geometry Using Average Shading Gradients [pdf] - Tobias Plotz, Stefan Roth
  • An Adaptive Data Representation for Robust Point-Set Registration and Merging [pdf] - Dylan Campbell, Lars Petersson
  • Globally Optimal 2D-3D Registration From Points or Lines Without Correspondences [pdf] - Mark Brown, David Windridge, Jean-Yves Guillemaut
Local Convolutional Features With Unsupervised Training for Image Retrieval [pdf]
Mattis Paulin, Matthijs Douze, Zaid Harchaoui, Julien Mairal, Florent Perronin, Cordelia Schmid

Abstract: Patch-level descriptors underlie several important computer vision tasks, such as stereo-matching or contentbased image retrieval. We introduce a deep convolutional architecture that yields patch-level descriptors, as an alternative to the popular SIFT descriptor for image retrieval. The proposed family of descriptors, called PatchCKN, adapt the recently introduced Convolutional Kernel Network (CKN), an unsupervised framework to learn convolutional architectures. We present a comparison framework to benchmark current deep convolutional approaches along with Patch-CKN for both patch and image retrieval, including our novel RomePatches dataset. Patch-CKN descriptors yield competitive results compared to supervised CNN alternatives on patch and image retrieval. Figure 1. Proposed approach: interest regions are first extracted from images; then the neighborhood of each point is affine- and orientation-normalized to sample a patch; finally, deep convolutional nets are used to get a descriptor for each patch.
Similar papers:
  • Unsupervised Learning of Visual Representations Using Videos [pdf] - Xiaolong Wang, Abhinav Gupta
  • Unsupervised Visual Representation Learning by Context Prediction [pdf] - Carl Doersch, Abhinav Gupta, Alexei A. Efros
  • Aggregating Local Deep Features for Image Retrieval [pdf] - Artem Babenko, Victor Lempitsky
  • RIDE: Reversal Invariant Descriptor Enhancement [pdf] - Lingxi Xie, Jingdong Wang, Weiyao Lin, Bo Zhang, Qi Tian
  • Discriminative Learning of Deep Convolutional Feature Point Descriptors [pdf] - Edgar Simo-Serra, Eduard Trulls, Luis Ferraz, Iasonas Kokkinos, Pascal Fua, Francesc Moreno-Noguer
Learning Where to Position Parts in 3D [pdf]
Marco Pedersoli, Tinne Tuytelaars

Abstract: A common issue in deformable object detection is finding a good way to position the parts. This issue is even more outspoken when considering detection and pose estimation for 3D objects, where parts should be placed in a threedimensional space. Some methods extract the 3D shape of the object from 3D CAD models. This limits their applicability to categories for which such models are available. Others represent the object with a pre-defined and simple shape (e.g. a cuboid). This extends the applicability of the model, but in many cases the pre-defined shape is too simple to properly represent the object in 3D. In this paper we propose a new method for the detection and pose estimation of 3D objects, that does not use any 3D CAD model or other 3D information. Starting from a simple and general 3D shape, we learn in a weakly supervised manner the 3D part locations that best fit the training data. As this method builds on a iterative estimation of the part locations, we introduce several speedups to make the method fast enough for practical experiments. We evaluate our model for the detection and pose estimation of faces and cars. Our method obtains results comparable with the state of the art, it is faster than most of the other approaches and does not need any additional 3D information.
Similar papers:
  • Single Image Pop-Up From Discriminatively Learned Parts [pdf] - Menglong Zhu, Xiaowei Zhou, Kostas Daniilidis
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • Scene-Domain Active Part Models for Object Representation [pdf] - Zhou Ren, Chaohui Wang, Alan L. Yuille
  • Pose Induction for Novel Object Categories [pdf] - Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • A Novel Representation of Parts for Accurate 3D Object Detection and Tracking in Monocular Images [pdf] - Alberto Crivellaro, Mahdi Rad, Yannick Verdie, Kwang Moo Yi, Pascal Fua, Vincent Lepetit
Learning Deep Object Detectors From 3D Models [pdf]
Xingchao Peng, Baochen Sun, Karim Ali, Kate Saenko

Abstract: Crowdsourced 3D CAD models are becoming easily accessible online, and can potentially generate an infinite number of training images for almost any object category. We show that augmenting the training data of contemporary Deep Convolutional Neural Net (DCNN) models with such synthetic data can be effective, especially when real training data is limited or not well matched to the target domain. Most freely available CAD models capture 3D shape but are often missing other low level cues, such as realistic object texture, pose, or background. In a detailed analysis, we use synthetic CAD-rendered images to probe the ability of DCNN to learn without these cues, with surprising findings. In particular, we show that when the DCNN is fine-tuned on the target detection task, it exhibits a large degree of invariance to missing low-level cues, but, when pretrained on generic ImageNet classification, it learns better when the low-level cues are simulated. We show that our synthetic DCNN training approach significantly outperforms previous methods on the PASCAL VOC2007 dataset when learning in the few-shot scenario and improves performance in a domain shift scenario on the Office benchmark.
Similar papers:
  • Webly Supervised Learning of Convolutional Networks [pdf] - Xinlei Chen, Abhinav Gupta
  • Simultaneous Deep Transfer Across Domains and Tasks [pdf] - Eric Tzeng, Judy Hoffman, Trevor Darrell, Kate Saenko
  • Render for CNN: Viewpoint Estimation in Images Using CNNs Trained With Rendered 3D Model Views [pdf] - Hao Su, Charles R. Qi, Yangyan Li, Leonidas J. Guibas
  • Example-Based Modeling of Facial Texture From Deficient Data [pdf] - Arnaud Dessein, William A. P. Smith, Richard C. Wilson, Edwin R. Hancock
  • Semantic Pose Using Deep Networks Trained on Synthetic RGB-D [pdf] - Jeremie Papon, Markus Schoeler
PIEFA: Personalized Incremental and Ensemble Face Alignment [pdf]
Xi Peng, Shaoting Zhang, Yu Yang, Dimitris N. Metaxas

Abstract: Face alignment, especially on real-time or large-scale sequential images, is a challenging task with broad applications. Both generic and joint alignment approaches have been proposed with varying degrees of success. However, many generic methods are heavily sensitive to initializations and usually rely on offline-trained static models, which limit their performance on sequential images with extensive variations. On the other hand, joint methods are restricted to offline applications, since they require all frames to conduct batch alignment. To address these limitations, we propose to exploit incremental learning for personalized ensemble alignment. We sample multiple initial shapes to achieve image congealing within one frame, which enables us to incrementally conduct ensemble alignment by group-sparse regularized rank minimization. At the same time, personalized modeling is obtained by subspace adaptation under the same incremental framework, while correction strategy is used to alleviate model drifting. Experimental results on multiple controlled and in-the-wild databases demonstrate the superior performance of our approach compared with state-of-the-arts in terms of fitting accuracy and efficiency.
Similar papers:
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
  • Robust Statistical Face Frontalization [pdf] - Christos Sagonas, Yannis Panagakis, Stefanos Zafeiriou, Maja Pantic
  • Personalized Age Progression With Aging Dictionary [pdf] - Xiangbo Shu, Jinhui Tang, Hanjiang Lai, Luoqi Liu, Shuicheng Yan
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
Fully Connected Object Proposals for Video Segmentation [pdf]
Federico Perazzi, Oliver Wang, Markus Gross, Alexander Sorkine-Hornung

Abstract: We present a novel approach to video segmentation using multiple object proposals. The problem is formulated as a minimization of a novel energy function defined over a fully connected graph of object proposals. Our model combines appearance with long-range point tracks, which is key to ensure robustness with respect to fast motion and occlusions over longer video sequences. As opposed to previous approaches based on object proposals, we do not seek the best per-frame object hypotheses to perform the segmentation. Instead, we combine multiple, potentially imperfect proposals to improve overall segmentation accuracy and ensure robustness to outliers. Overall, the basic algorithm consists of three steps. First, we generate a very large number of object proposals for each video frame using existing techniques. Next, we perform an SVM-based pruning step to retain only high quality proposals with sufficiently discriminative power. Finally, we determine the fore- and background classification by solving for the maximum a posteriori of a fully connected conditional random field, defined using our novel energy function. Experimental results on a well established dataset demonstrate that our method compares favorably to several recent state-of-the-art approaches.
Similar papers:
  • Cascaded Sparse Spatial Bins for Efficient and Effective Generic Object Detection [pdf] - David Novotny, Jiri Matas
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • Box Aggregation for Proposal Decimation: Last Mile of Object Detection [pdf] - Shu Liu, Cewu Lu, Jiaya Jia
Flowing ConvNets for Human Pose Estimation in Videos [pdf]
Tomas Pfister, James Charles, Andrew Zisserman

Abstract: The objective of this work is human pose estimation in videos, where multiple frames are available. We investigate a ConvNet architecture that is able to benefit from temporal context by combining information across the multiple frames using optical flow. To this end we propose a network architecture with the following novelties: (i) a deeper network than previously investigated for regressing heatmaps; (ii) spatial fusion layers that learn an implicit spatial model; (iii) optical flow is used to align heatmap predictions from neighbouring frames; and (iv) a final parametric pooling layer which learns to combine the aligned heatmaps into a pooled confidence map. We show that this architecture outperforms a number of others, including one that uses optical flow solely at the input layers, one that regresses joint coordinates directly, and one that predicts heatmaps without spatial fusion. The new architecture outperforms the state of the art by a large margin on three video pose estimation datasets, including the very challenging Poses in the Wild dataset, and outperforms other deep methods that dont use a graphical model on the single-image FLIC benchmark (and also [5, 35] in the high precision region).
Similar papers:
  • PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization [pdf] - Alex Kendall, Matthew Grimes, Roberto Cipolla
  • Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf] - Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang
  • Recurrent Network Models for Human Dynamics [pdf] - Katerina Fragkiadaki, Sergey Levine, Panna Felsen, Jitendra Malik
  • P-CNN: Pose-Based CNN Features for Action Recognition [pdf] - Guilhem Cheron, Ivan Laptev, Cordelia Schmid
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
COUNT Forest: CO-Voting Uncertain Number of Targets Using Random Forest for Crowd Density Estimation [pdf]
Viet-Quoc Pham, Tatsuo Kozakaya, Osamu Yamaguchi, Ryuzo Okada

Abstract: This paper presents a patch-based approach for crowd density estimation in public scenes. We formulate the problem of estimating density in a structured learning framework applied to random decision forests. Our approach learns the mapping between patch features and relative locations of all objects inside each patch, which contribute to generate the patch density map through Gaussian kernel density estimation. We build the forest in a coarse-to-fine manner with two split node layers, and further propose a crowdedness prior and an effective forest reduction method to improve the estimation accuracy and speed. Moreover, we introduce a semi-automatic training method to learn the estimator for a specific scene. We achieved state-of-the-art results on the public Mall dataset and UCSD dataset, and also proposed two potential applications in traffic counts and scene understanding with promising results.
Similar papers:
  • Naive Bayes Super-Resolution Forest [pdf] - Jordi Salvador, Eduardo Perez-Pellitero
  • Deep Neural Decision Forests [pdf] - Peter Kontschieder, Madalina Fiterau, Antonio Criminisi, Samuel Rota Bulo
  • Live Repetition Counting [pdf] - Ofir Levy, Lior Wolf
  • Pedestrian Travel Time Estimation in Crowded Scenes [pdf] - Shuai Yi, Hongsheng Li, Xiaogang Wang
  • Bayesian Model Adaptation for Crowd Counts [pdf] - Bo Liu, Nuno Vasconcelos
Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf]
Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould

Abstract: Although there has been remarkable progress in 2D image semantic segmentation, 3D scene semantic labelling has not received as much attention or achievement (though some notable exceptions exist, e.g., [12, 8]). Unlike 2D images which capture specific views, 3D reconstructed point clouds cover the whole scene with a large number of things and stuff, making label prediction more challenging. Yet 3D data inherently carries rich contextual information beneficial for the label prediction task if successfully exploited. Koppula et al. [12] proposed a graphical model with sophisticated pairwise potentials to encode contextual relations between different objects for 3D indoor scene understanding. The model parameters are learned from training data using a Structured Support Vector Machine (SSVM) approach. Kahler and Reid [8] learn similar potential functions using Decision Tree Fields (DTF) [19] and Regression Tree Fields (RTF) [7]. The tree based method [8] is more efficient and yields similar segmentation accuracies. Although initial promising results have been reported, both of these models restrict themselves to pairwise potential functions limiting their ability to encode complex geometrical and topological characteristics of 3D scenes. We propose to overcome the above-mentioned limitations of previous work by introducing a novel higher-order model for semantic 3D indoor scene labelling. Our higherorder potentials can capture much more structural information embedded in the sc
Similar papers:
  • Semantic Segmentation of RGBD Images With Mutex Constraints [pdf] - Zhuo Deng, Sinisa Todorovic, Longin Jan Latecki
  • Semantic Segmentation With Object Clique Potential [pdf] - Xiaojuan Qi, Jianping Shi, Shu Liu, Renjie Liao, Jiaya Jia
  • Semantic Image Segmentation via Deep Parsing Network [pdf] - Ziwei Liu, Xiaoxiao Li, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Context-Aware CNNs for Person Head Detection [pdf] - Tuan-Hung Vu, Anton Osokin, Ivan Laptev
  • Parsimonious Labeling [pdf] - Puneet K. Dokania, M. Pawan Kumar
Registering Images to Untextured Geometry Using Average Shading Gradients [pdf]
Tobias Plotz, Stefan Roth

Abstract: Many existing approaches for image-to-geometry registration assume that either a textured 3D model or a good initial guess of the 3D pose is available to bootstrap the registration process. In this paper we consider the registration of photographs to 3D models even when no texture information is available. This is very challenging as we cannot rely on texture gradients, and even shading gradients are hard to estimate since the lighting conditions are unknown. To that end, we propose average shading gradients, a rendering technique that estimates the average gradient magnitude over all lighting directions under Lambertian shading. We use this gradient representation as the building block of a registration pipeline based on matching sparse features. To cope with inevitable false matches due to the missing texture information and to increase robustness, the pose of the 3D model is estimated in two stages. Coarse pose hypotheses are first obtained from a single correct match each, subsequently refined using SIFT flow, and finally verified. We apply our algorithm to registering images of real-world objects to untextured 3D meshes of limited accuracy.
Similar papers:
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • An Adaptive Data Representation for Robust Point-Set Registration and Merging [pdf] - Dylan Campbell, Lars Petersson
  • Direct Intrinsics: Learning Albedo-Shading Decomposition by Convolutional Regression [pdf] - Takuya Narihira, Michael Maire, Stella X. Yu
  • See the Difference: Direct Pre-Image Reconstruction and Pose Estimation by Differentiating HOG [pdf] - Wei-Chen Chiu, Mario Fritz
  • Intrinsic Decomposition of Image Sequences From Local Temporal Variations [pdf] - Pierre-Yves Laffont, Jean-Charles Bazin
Flickr30k Entities: Collecting Region-to-Phrase Correspondences for Richer Image-to-Sentence Models [pdf]
Bryan A. Plummer, Liwei Wang, Chris M. Cervantes, Juan C. Caicedo, Julia Hockenmaier, Svetlana Lazebnik

Abstract: The Flickr30k dataset has become a standard benchmark for sentence-based image description. This paper presents Flickr30k Entities, which augments the 158k captions from Flickr30k with 244k coreference chains linking mentions of the same entities in images, as well as 276k manually annotated bounding boxes corresponding to each entity. Such annotation is essential for continued progress in automatic image description and grounded language understanding. We present experiments demonstrating the usefulness of our annotations for text-to-image reference resolution, or the task of localizing textual entity mentions in an image, and for bidirectional image-sentence retrieval. These experiments confirm that we can further improve the accuracy of state-of-the-art retrieval methods by training with explicit region-to-phrase correspondence, but at the same time, they show that accurately inferring this correspondence given an image and a caption remains really challenging.
Similar papers:
  • Visual Phrases for Exemplar Face Detection [pdf] - Vijay Kumar, Anoop Namboodiri, C. V. Jawahar
  • Multimodal Convolutional Neural Networks for Matching Image and Sentence [pdf] - Lin Ma, Zhengdong Lu, Lifeng Shang, Hang Li
  • Guiding the Long-Short Term Memory Model for Image Caption Generation [pdf] - Xu Jia, Efstratios Gavves, Basura Fernando, Tinne Tuytelaars
  • Segment-Phrase Table for Semantic Segmentation, Visual Entailment and Paraphrasing [pdf] - Hamid Izadinia, Fereshteh Sadeghi, Santosh K. Divvala, Hannaneh Hajishirzi, Yejin Choi, Ali Farhadi
  • Common Subspace for Model and Similarity: Phrase Learning for Caption Generation From Images [pdf] - Yoshitaka Ushiku, Masataka Yamaguchi, Yusuke Mukuta, Tatsuya Harada
Boosting Object Proposals: From Pascal to COCO [pdf]
Jordi Pont-Tuset, Luc Van Gool

Abstract: Computer vision in general, and object proposals in particular, are nowadays strongly influenced by the databases on which researchers evaluate the performance of their algorithms. This paper studies the transition from the Pascal Visual Object Challenge dataset, which has been the benchmark of reference for the last years, to the updated, bigger, and more challenging Microsoft Common Objects in Context. We first review and deeply analyze the new challenges, and opportunities, that this database presents. We then survey the current state of the art in object proposals and evaluate it focusing on how it generalizes to the new dataset. In sight of these results, we propose various lines of research to take advantage of the new benchmark and improve the techniques. We explore one of these lines, which leads to an improvement over the state of the art of +5.2%.
Similar papers:
  • DeepProposal: Hunting Objects by Cascading Deep Convolutional Layers [pdf] - Amir Ghodrati, Ali Diba, Marco Pedersoli, Tinne Tuytelaars, Luc Van Gool
  • The Middle Child Problem: Revisiting Parametric Min-Cut and Seeds for Object Proposals [pdf] - Ahmad Humayun, Fuxin Li, James M. Rehg
  • Box Aggregation for Proposal Decimation: Last Mile of Object Detection [pdf] - Shu Liu, Cewu Lu, Jiaya Jia
  • Fully Connected Object Proposals for Video Segmentation [pdf] - Federico Perazzi, Oliver Wang, Markus Gross, Alexander Sorkine-Hornung
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf]
Niloufar Pourian, S. Karthikeyan, B.S. Manjunath

Abstract: We present a weakly-supervised approach to semantic segmentation. The goal is to assign pixel-level labels given only partial information, for example, image-level labels. This is an important problem in many application scenarios where it is difficult to get accurate segmentation or not feasible to obtain detailed annotations. The proposed approach starts with an initial coarse segmentation, followed by a spectral clustering approach that groups related image parts into communities. A community-driven graph is then constructed that captures spatial and feature relationships between communities while a label graph captures correlations between image labels. Finally, mapping the image level labels to appropriate communities is formulated as a convex optimization problem. The proposed approach does not require location information for image level labels and can be trained using partially labeled datasets. Compared to the state-of-the-art weakly supervised approaches, we achieve a significant performance improvement of 9% on MSRC-21 dataset and 11% on LabelMe dataset, while being more than 300 times faster.
Similar papers:
  • Weakly- and Semi-Supervised Learning of a Deep Convolutional Network for Semantic Image Segmentation [pdf] - George Papandreou, Liang-Chieh Chen, Kevin P. Murphy, Alan L. Yuille
  • Augmenting Strong Supervision Using Web Data for Fine-Grained Categorization [pdf] - Zhe Xu, Shaoli Huang, Ya Zhang, Dacheng Tao
  • Probabilistic Label Relation Graphs With Ising Models [pdf] - Nan Ding, Jia Deng, Kevin P. Murphy, Hartmut Neven
  • Joint Object and Part Segmentation Using Deep Learned Potentials [pdf] - Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille
  • ML-MG: Multi-Label Learning With Missing Labels Using a Mixed Graph [pdf] - Baoyuan Wu, Siwei Lyu, Bernard Ghanem
Attribute-Graph: A Graph Based Approach to Image Ranking [pdf]
Nikita Prabhu, R. Venkatesh Babu

Abstract: We propose a novel image representation, termed Attribute-Graph, to rank images by their semantic similarity to a given query image. An Attribute-Graph is an undirected fully connected graph, incorporating both local and global image characteristics. The graph nodes characterise objects as well as the overall scene context using mid-level semantic attributes, while the edges capture the object topology. We demonstrate the effectiveness of Attribute-Graphs by applying them to the problem of image ranking. We benchmark the performance of our algorithm on the rPascal and rImageNet datasets, which we have created in order to evaluate the ranking performance on complex queries containing multiple objects. Our experimental evaluation shows that modelling images as Attribute-Graphs results in improved ranking performance over existing techniques.
Similar papers:
  • Discrete Tabu Search for Graph Matching [pdf] - Kamil Adamczewski, Yumin Suh, Kyoung Mu Lee
  • Learning Deep Representation With Large-Scale Attributes [pdf] - Wanli Ouyang, Hongyang Li, Xingyu Zeng, Xiaogang Wang
  • Learning Query and Image Similarities With Ranking Canonical Correlation Analysis [pdf] - Ting Yao, Tao Mei, Chong-Wah Ngo
  • Multi-Task Learning With Low Rank Attribute Embedding for Person Re-Identification [pdf] - Chi Su, Fan Yang, Shiliang Zhang, Qi Tian, Larry S. Davis, Wen Gao
  • Mining And-Or Graphs for Graph Matching and Object Discovery [pdf] - Quanshi Zhang, Ying Nian Wu, Song-Chun Zhu
MAP Disparity Estimation Using Hidden Markov Trees [pdf]
Eric T. Psota, Jedrzej Kowalczuk, Mateusz Mittek, Lance C. Perez

Abstract: A new method is introduced for stereo matching that operates on minimum spanning trees (MSTs) generated from the images. Disparity maps are represented as a collection of hidden states on MSTs, and each MST is modeled as a hidden Markov tree. An efficient recursive message-passing scheme designed to operate on hidden Markov trees, known as the upward-downward algorithm, is used to compute the maximum a posteriori (MAP) disparity estimate at each pixel. The messages processed by the upward-downward algorithm involve two types of probabilities: the probability of a pixel having a particular disparity given a set of per-pixel matching costs, and the probability of a disparity transition between a pair of connected pixels given their similarity. The distributions of these probabilities are modeled from a collection of images with ground truth disparities. Performance evaluation using the Middlebury stereo benchmark version 3 demonstrates that the proposed method ranks second and third in terms of overall accuracy when evaluated on the training and test image sets, respectively.
Similar papers:
  • MeshStereo: A Global Stereo Model With Mesh Alignment Regularization for View Interpolation [pdf] - Chi Zhang, Zhiwei Li, Yanhua Cheng, Rui Cai, Hongyang Chao, Yong Rui
  • The HCI Stereo Metrics: Geometry-Aware Performance Analysis of Stereo Algorithms [pdf] - Katrin Honauer, Lena Maier-Hein, Daniel Kondermann
  • Blur-Aware Disparity Estimation From Defocus Stereo Images [pdf] - Ching-Hui Chen, Hui Zhou, Timo Ahonen
  • Depth Map Estimation and Colorization of Anaglyph Images Using Local Color Prior and Reverse Intensity Distribution [pdf] - Williem, Ramesh Raskar, In Kyu Park
  • A Deep Visual Correspondence Embedding Model for Stereo Matching Costs [pdf] - Zhuoyuan Chen, Xun Sun, Liang Wang, Yinan Yu, Chang Huang
Rolling Shutter Super-Resolution [pdf]
Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan

Abstract: Classical multi-image super-resolution (SR) algorithms, designed for CCD cameras, assume that the motion among the images is global. But CMOS sensors that have increasingly started to replace their more expensive CCD counterparts in many applications do not respect this assumption if there is a motion of the camera relative to the scene during the exposure duration of an image because of the rowwise acquisition mechanism. In this paper, we study the hitherto unexplored topic of multi-image SR in CMOS cameras. We initially develop an SR observation model that accounts for the row-wise distortions called the rolling shutter (RS) effect observed in images captured using nonstationary CMOS cameras. We then propose a unified RSSR framework to obtain an RS-free high-resolution image (and the row-wise motion) from distorted low-resolution images. We demonstrate the efficacy of the proposed scheme using synthetic data as well as real images captured using a hand-held CMOS camera. Quantitative and qualitative assessments reveal that our method significantly advances the state-of-the-art.
Similar papers:
  • Model-Based Tracking at 300Hz Using Raw Time-of-Flight Observations [pdf] - Jan Stuhmer, Sebastian Nowozin, Andrew Fitzgibbon, Richard Szeliski, Travis Perry, Sunil Acharya, Daniel Cremers, Jamie Shotton
  • An Efficient Minimal Solution for Multi-Camera Motion [pdf] - Jonathan Ventura, Clemens Arth, Vincent Lepetit
  • Integrating Dashcam Views Through Inter-Video Mapping [pdf] - Hsin-I Chen, Yi-Ling Chen, Wei-Tse Lee, Fan Wang, Bing-Yu Chen
  • Dense Continuous-Time Tracking and Mapping With Rolling Shutter RGB-D Cameras [pdf] - Christian Kerl, Jorg Stuckler, Daniel Cremers
  • SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf] - Kensei Jo, Mohit Gupta, Shree K. Nayar
Unsupervised Tube Extraction Using Transductive Learning and Dense Trajectories [pdf]
Mihai Marian Puscas, Enver Sangineto, Dubravko Culibrk, Nicu Sebe

Abstract: We address the problem of automatic extraction of foreground objects from videos. The goal is to provide a method for unsupervised collection of samples which can be further used for object detection training without any human intervention. We use the well known Selective Search approach to produce an initial still-image based segmentation of the video frames. This initial set of proposals is pruned and temporally extended using optical flow and transductive learning. Specifically, we propose to use Dense Trajectories in order to robustly match and track candidate boxes over different frames. The obtained box tracks are used to collect samples for unsupervised training of track-specific detectors. Finally, the detectors are run on the videos to extract the final tubes. The combination of appearance-based static objectness (Selective Search), motion information (Dense Trajectories) and transductive learning (detectors are forced to overfit on the unsupervised data used for training) makes the proposed approach extremely robust. We outperform state-of-the-art systems by a large margin on common benchmarks used for tube proposal evaluation.
Similar papers:
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
  • Face Flow [pdf] - Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou
  • Flow Fields: Dense Correspondence Fields for Highly Accurate Large Displacement Optical Flow Estimation [pdf] - Christian Bailer, Bertram Taetz, Didier Stricker
  • Dense Optical Flow Prediction From a Static Image [pdf] - Jacob Walker, Abhinav Gupta, Martial Hebert
Semantic Segmentation With Object Clique Potential [pdf]
Xiaojuan Qi, Jianping Shi, Shu Liu, Renjie Liao, Jiaya Jia

Abstract: We propose an object clique potential for semantic segmentation. Our object clique potential addresses the misclassified object-part issues arising in solutions based on fully-convolutional networks. Our object clique set, compared to that yielded from segment-proposal-based approaches, is with a significantly smaller size, making our method consume notably less computation. Regarding system design and model formation, our object clique potential can be regarded as a functional complement to local-appearance-based CRF models and works in synergy with these effective approaches for further performance improvement. Extensive experiments verify our method. (a) Input (b) Segment proposal result [3] (c) Input (d) FCN result [24] (e) Input (f) CRF after FCN result [2]
Similar papers:
  • Parsimonious Labeling [pdf] - Puneet K. Dokania, M. Pawan Kumar
  • Conditional Random Fields as Recurrent Neural Networks [pdf] - Shuai Zheng, Sadeep Jayasumana, Bernardino Romera-Paredes, Vibhav Vineet, Zhizhong Su, Dalong Du, Chang Huang, Philip H. S. Torr
  • Joint Object and Part Segmentation Using Deep Learned Potentials [pdf] - Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille
  • Semantic Image Segmentation via Deep Parsing Network [pdf] - Ziwei Liu, Xiaoxiao Li, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
Frequency-Based Environment Matting by Compressive Sensing [pdf]
Yiming Qian, Minglun Gong, Yee-Hong Yang

Abstract: Extracting environment mattes using existing approaches often requires either thousands of captured images or a long processing time, or both. In this paper, we propose a novel approach to capturing and extracting the matte of a real scene effectively and efficiently. Grown out of the traditional frequency-based signal analysis, our approach can accurately locate contributing sources. By exploiting the recently developed compressive sensing theory, we simplify the data acquisition process of frequency-based environment matting. Incorporating phase information in a frequency signal into data acquisition further accelerates the matte extraction procedure. Compared with the stateof-the-art method, our approach achieves superior performance on both synthetic and real data, while consuming only a fraction of the processing time.
Similar papers:
  • Robust Heart Rate Measurement From Video Using Select Random Patches [pdf] - Antony Lam, Yoshinori Kuno
  • Depth Selective Camera: A Direct, On-Chip, Programmable Technique for Depth Selectivity in Photography [pdf] - Ryuichi Tadano, Adithya Kumar Pediredla, Ashok Veeraraghavan
  • Video Matting via Sparse and Low-Rank Representation [pdf] - Dongqing Zou, Xiaowu Chen, Guangying Cao, Xiaogang Wang
  • Image Matting With KL-Divergence Based Sparse Sampling [pdf] - Levent Karacan, Aykut Erdem, Erkut Erdem
  • 3D Surface Profilometry Using Phase Shifting of De Bruijn Pattern [pdf] - Matea Donlic, Tomislav Petkovic, Tomislav Pribanic
Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf]
Yuhui Quan, Yan Huang, Hui Ji

Abstract: Dynamic textures (DTs) are video sequences with stationary properties, which exhibit repetitive patterns over space and time. This paper aims at investigating the sparse coding based approach to characterizing local DT patterns for recognition. Owing to the high dimensionality of DT sequences, existing dictionary learning algorithms are not suitable for our purpose due to their high computational costs as well as poor scalability. To overcome these obstacles, we proposed a structured tensor dictionary learning method for sparse coding, which learns a dictionary structured with orthogonality and separability. The proposed method is very fast and more scalable to high-dimensional data than the existing ones. In addition, based on the proposed dictionary learning method, a DT descriptor is developed, which has better adaptivity, discriminability and scalability than the existing approaches. These advantages are demonstrated by the experiments on multiple datasets.
Similar papers:
  • Convolutional Sparse Coding for Image Super-Resolution [pdf] - Shuhang Gu, Wangmeng Zuo, Qi Xie, Deyu Meng, Xiangchu Feng, Lei Zhang
  • Person Re-Identification With Discriminatively Trained Viewpoint Invariant Dictionaries [pdf] - Srikrishna Karanam, Yang Li, Richard J. Radke
  • Variational Depth Superresolution Using Example-Based Edge Representations [pdf] - David Ferstl, Matthias Ruther, Horst Bischof
  • Linearization to Nonlinear Learning for Visual Tracking [pdf] - Bo Ma, Hongwei Hu, Jianbing Shen, Yuping Zhang, Fatih Porikli
  • Conformal and Low-Rank Sparse Representation for Image Restoration [pdf] - Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng
Learning Temporal Embeddings for Complex Video Analysis [pdf]
Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei

Abstract: Figure 1. The temporal context of a video frame is crucial in determining its true semantic meaning. For instance, consider the above example where the embeddings of different semantic classes are shown in different colors. The middle frame from the two wedding videos correspond to visually dissimilar classes of church ceremony and court ceremony. However, by observing the similarity in their temporal contexts we expect them to be semantically closer. Our work leverages such powerful temporal context to learn semantically rich embeddings.
Similar papers:
  • Video Restoration Against Yin-Yang Phasing [pdf] - Xiaolin Wu, Zhenhao Li, Xiaowei Deng
  • Storyline Representation of Egocentric Videos With an Applications to Story-Based Search [pdf] - Bo Xiong, Gunhee Kim, Leonid Sigal
  • Temporal Perception and Prediction in Ego-Centric Video [pdf] - Yipin Zhou, Tamara L. Berg
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
Multi-Label Cross-Modal Retrieval [pdf]
Viresh Ranjan, Nikhil Rasiwasia, C. V. Jawahar

Abstract: In this work, we address the problem of cross-modal retrieval in presence of multi-label annotations. In particular, we introduce multi-label Canonical Correlation Analysis (ml-CCA), an extension of CCA, for learning shared subspaces taking into account the high level semantic information in the form of multi-label annotations. Unlike CCA, ml-CCA does not rely on explicit pairings between the modalities, instead it uses the multi-label information to establish correspondences. This results in a discriminative subspace which is better suited for cross-modal retrieval tasks. We also present Fast ml-CCA, a computationally efficient version of ml-CCA, which is able to handle large scale datasets. We show the efficacy of our approach by conducting extensive cross-modal retrieval experiments on three standard benchmark datasets. The results show that the proposed approach achieves state-of-the-art retrieval performance on the three datasets.
Similar papers:
  • Learning Query and Image Similarities With Ranking Canonical Correlation Analysis [pdf] - Ting Yao, Tao Mei, Chong-Wah Ngo
  • ML-MG: Multi-Label Learning With Missing Labels Using a Mixed Graph [pdf] - Baoyuan Wu, Siwei Lyu, Bernard Ghanem
  • Similarity Gaussian Process Latent Variable Model for Multi-Modal Data Analysis [pdf] - Guoli Song, Shuhui Wang, Qingming Huang, Qi Tian
  • Cross-Domain Image Retrieval With a Dual Attribute-Aware Ranking Network [pdf] - Junshi Huang, Rogerio S. Feris, Qiang Chen, Shuicheng Yan
  • Localize Me Anywhere, Anytime: A Multi-Task Point-Retrieval Approach [pdf] - Guoyu Lu, Yan Yan, Li Ren, Jingkuan Song, Nicu Sebe, Chandra Kambhamettu
Object Detection Using Generalization and Efficiency Balanced Co-Occurrence Features [pdf]
Haoyu Ren, Ze-Nian Li

Abstract: In this paper, we propose a high-accuracy object detector based on co-occurrence features. Firstly, we introduce three kinds of local co-occurrence features constructed by the traditional Haar, LBP, and HOG respectively. Then the boosted detectors are learned, where each weak classifier corresponds to a local image region with a co-occurrence feature. In addition, we propose a Generalization and Efficiency Balanced (GEB) framework for boosting training. In the feature selection procedure, the discrimination ability, the generalization power, and the computation cost of the candidate features are all evaluated for decision. As a result, the boosted detector achieves both high accuracy and good efficiency. It also shows the performance competitive with the state-of-the-art methods for pedestrian detection and general object detection tasks. Object detection is an indispensable technology in many applications such as artificial intelligence, multimedia systems, and video surveillance. The major problem is that the object appearances vary greatly because of different illuminations, view points, poses, and the presence of occlusions. This has motivated the invention of various approaches. Among them, a commonly used paradigm is to train a boosted classifier based on local features [14][26]. For example, Viola et al. [26] used AdaBoost algorithm to train a cascade classifier based on the Haar feature. Zhang et al. [36] proposed an improved version of the Haar feature based on up-ri
Similar papers:
  • Geometry-Aware Deep Transform [pdf] - Jiaji Huang, Qiang Qiu, Robert Calderbank, Guillermo Sapiro
  • Learning a Descriptor-Specific 3D Keypoint Detector [pdf] - Samuele Salti, Federico Tombari, Riccardo Spezialetti, Luigi Di Stefano
  • Convolutional Channel Features [pdf] - Bin Yang, Junjie Yan, Zhen Lei, Stan Z. Li
  • Deep Learning Strong Parts for Pedestrian Detection [pdf] - Yonglong Tian, Ping Luo, Xiaogang Wang, Xiaoou Tang
  • Learning Complexity-Aware Cascades for Deep Pedestrian Detection [pdf] - Zhaowei Cai, Mohammad Saberian, Nuno Vasconcelos
Scene-Domain Active Part Models for Object Representation [pdf]
Zhou Ren, Chaohui Wang, Alan L. Yuille

Abstract: In this paper, we are interested in enhancing the expressivity and robustness of part-based models for object representation, in the common scenario where the training data are based on 2D images. To this end, we propose scenedomain active part models (SDAPM), which reconstruct and characterize the 3D geometric statistics between objects parts in 3D scene-domain by using 2D training data in the image-domain alone. And on top of this, we explicitly model and handle occlusions in SDAPM. Together with the developed learning and inference algorithms, such a model provides rich object descriptions, including 2D object and parts localization, 3D landmark shape and camera viewpoint, which offers an effective representation to various image understanding tasks, such as object and parts detection, 3D landmark shape and viewpoint estimation from images. Experiments on the above tasks show that SDAPM outperforms previous part-based models, and thus demonstrates the potential of the proposed technique.
Similar papers:
  • Unsupervised Generation of a Viewpoint Annotated Car Dataset From Videos [pdf] - Nima Sedaghat, Thomas Brox
  • Robust Facial Landmark Detection Under Significant Head Poses and Occlusion [pdf] - Yue Wu, Qiang Ji
  • Learning Where to Position Parts in 3D [pdf] - Marco Pedersoli, Tinne Tuytelaars
  • Render for CNN: Viewpoint Estimation in Images Using CNNs Trained With Rendered 3D Model Views [pdf] - Hao Su, Charles R. Qi, Yangyan Li, Leonidas J. Guibas
  • Single Image Pop-Up From Discriminatively Learned Parts [pdf] - Menglong Zhu, Xiaowei Zhou, Kostas Daniilidis
Joint Probabilistic Data Association Revisited [pdf]
Seyed Hamid Rezatofighi, Anton Milan, Zhen Zhang, Qinfeng Shi, Anthony Dick, Ian Reid

Abstract: 4 8 In this paper, we revisit the joint probabilistic data association (JPDA) technique and propose a novel solution based on recent developments in finding the m-best solutions to an integer linear program. The key advantage of this approach is that it makes JPDA computationally tractable in applications with high target and/or clutter density, such as spot tracking in fluorescence microscopy sequences and pedestrian tracking in surveillance footage. We also show that our JPDA algorithm embedded in a simple tracking framework is surprisingly competitive with state-of-the-art global tracking methods in these two applications, while needing considerably less processing time.
Similar papers:
  • A Versatile Learning-Based 3D Temporal Tracker: Scalable, Robust, Online [pdf] - David Joseph Tan, Federico Tombari, Slobodan Ilic, Nassir Navab
  • Multiple Hypothesis Tracking Revisited [pdf] - Chanho Kim, Fuxin Li, Arridhana Ciptadi, James M. Rehg
  • Learning to Divide and Conquer for Online Multi-Target Tracking [pdf] - Francesco Solera, Simone Calderara, Rita Cucchiara
  • Learning to Track: Online Multi-Object Tracking by Decision Making [pdf] - Yu Xiang, Alexandre Alahi, Silvio Savarese
  • Near-Online Multi-Target Tracking With Aggregated Local Flow Descriptor [pdf] - Wongun Choi
A Versatile Scene Model With Differentiable Visibility Applied to Generative Pose Estimation [pdf]
Helge Rhodin, Nadia Robertini, Christian Richardt, Hans-Peter Seidel, Christian Theobalt

Abstract: Generative reconstruction methods compute the 3D configuration (such as pose and/or geometry) of a shape by optimizing the overlap of the projected 3D shape model with images. Proper handling of occlusions is a big challenge, since the visibility function that indicates if a surface point is seen from a camera can often not be formulated in closed form, and is in general discrete and non-differentiable at occlusion boundaries. We present a new scene representation that enables an analytically differentiable closed-form formulation of surface visibility. In contrast to previous methods, this yields smooth, analytically differentiable, and efficient to optimize pose similarity energies with rigorous occlusion handling, fewer local minima, and experimentally verified improved convergence of numerical optimization. The underlying idea is a new image formation model that represents opaque objects by a translucent medium with a smooth Gaussian density distribution which turns visibility into a smooth phenomenon. We demonstrate the advantages of our versatile scene model in several generative pose estimation problems, namely marker-less multi-object pose estimation, marker-less human motion capture with few cameras, and image-based 3D geometry estimation. ically well-behaved, i.e. smooth. Many approaches already fail to satisfy the first condition and use similarity functions that cannot be expressed or differentiated analytically. This necessitates the use of computationally expen
Similar papers:
  • TransCut: Transparent Object Segmentation From a Light-Field Image [pdf] - Yichao Xu, Hajime Nagahara, Atsushi Shimada, Rin-ichiro Taniguchi
  • Dense Image Registration and Deformable Surface Reconstruction in Presence of Occlusions and Minimal Texture [pdf] - Dat Tien Ngo, Sanghyuk Park, Anne Jorstad, Alberto Crivellaro, Chang D. Yoo, Pascal Fua
  • Beyond Tree Structure Models: A New Occlusion Aware Graphical Model for Human Pose Estimation [pdf] - Lianrui Fu, Junge Zhang, Kaiqi Huang
  • Robust Facial Landmark Detection Under Significant Head Poses and Occlusion [pdf] - Yue Wu, Qiang Ji
  • On the Visibility of Point Clouds [pdf] - Sagi Katz, Ayellet Tal
Uncovering Interactions and Interactors: Joint Estimation of Head, Body Orientation and F-Formations From Surveillance Videos [pdf]
Elisa Ricci, Jagannadan Varadarajan, Ramanathan Subramanian, Samuel Rota Bulo, Narendra Ahuja, Oswald Lanz

Abstract: We present a novel approach for jointly estimating targets head, body orientations and conversational groups called F-formations from a distant social scene (e.g., a cocktail party captured by surveillance cameras). Differing from related works that have (i) coupled head and body pose learning by exploiting the limited range of orientations that the two can jointly take, or (ii) determined F-formations based on the mutual head (but not body) orientations of interactors, we present a unified framework to jointly infer both (i) and (ii). Apart from exploiting spatial and orientation relationships, we also integrate cues pertaining to temporal consistency and occlusions, which are beneficial while handling low-resolution data under surveillance settings. Efficacy of the joint inference framework reflects via increased head, body pose and F-formation estimation accuracy over the state-of-the-art, as confirmed by extensive experiments on two social datasets.
Similar papers:
  • Fast and Accurate Head Pose Estimation via Random Projection Forests [pdf] - Donghoon Lee, Ming-Hsuan Yang, Songhwai Oh
  • Detailed Full-Body Reconstructions of Moving People From Monocular RGB-D Sequences [pdf] - Federica Bogo, Michael J. Black, Matthew Loper, Javier Romero
  • Robust Model-Based 3D Head Pose Estimation [pdf] - Gregory P. Meyer, Shalini Gupta, Iuri Frosio, Dikpal Reddy, Jan Kautz
  • Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf] - Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang
  • Human Pose Estimation in Videos [pdf] - Dong Zhang, Mubarak Shah
Conditioned Regression Models for Non-Blind Single Image Super-Resolution [pdf]
Gernot Riegler, Samuel Schulter, Matthias Ruther, Horst Bischof

Abstract: Single image super-resolution is an important task in the field of computer vision and finds many practical applications. Current state-of-the-art methods typically rely on machine learning algorithms to infer a mapping from low- to high-resolution images. These methods use a single fixed blur kernel during training and, consequently, assume the exact same kernel underlying the image formation process for all test images. However, this setting is not realistic for practical applications, because the blur is typically different for each test image. In this paper, we loosen this restrictive constraint and propose conditioned regression models (including convolutional neural networks and random forests) that can effectively exploit the additional kernel information during both, training and inference. This allows for training a single model, while previous methods need to be re-trained for every blur kernel individually to achieve good results, which we demonstrate in our evaluations. We also empirically show that the proposed conditioned regression models (i) can effectively handle scenarios where the blur kernel is different for each image and (ii) outperform related approaches trained for only a single kernel.
Similar papers:
  • Low Dimensional Explicit Feature Maps [pdf] - Ondrej Chum
  • Accurate Camera Calibration Robust to Defocus Using a Smartphone [pdf] - Hyowon Ha, Yunsu Bok, Kyungdon Joo, Jiyoung Jung, In So Kweon
  • Class-Specific Image Deblurring [pdf] - Saeed Anwar, Cong Phuoc Huynh, Fatih Porikli
  • Blur-Aware Disparity Estimation From Defocus Stereo Images [pdf] - Ching-Hui Chen, Hui Zhou, Timo Ahonen
  • Self-Calibration of Optical Lenses [pdf] - Michael Hirsch, Bernhard Scholkopf
Improving Ferns Ensembles by Sparsifying and Quantising Posterior Probabilities [pdf]
Antonio L. Rodriguez, Vitor Sequeira

Abstract: Ferns ensembles offer an accurate and efficient multiclass non-linear classification, commonly at the expense of consuming a large amount of memory. We introduce a two-fold contribution that produces large reductions in their memory consumption. First, an efficient L0 regularised cost optimisation finds a sparse representation of the posterior probabilities in the ensemble by discarding elements with zero contribution to valid responses in the training samples. As a by-product this can produce a prediction accuracy gain that, if required, can be traded for further reductions in memory size and prediction time. Secondly, posterior probabilities are quantised and stored in a memoryfriendly sparse data structure. We reported a minimum of 75% memory reduction for different types of classification problems using generative and discriminative ferns ensembles, without increasing prediction time or classification error. For image patch recognition our proposal produced a 90% memory reduction, and improved in several percentage points the prediction accuracy.
Similar papers:
  • Deep Neural Decision Forests [pdf] - Peter Kontschieder, Madalina Fiterau, Antonio Criminisi, Samuel Rota Bulo
  • Understanding and Predicting Image Memorability at a Large Scale [pdf] - Aditya Khosla, Akhil S. Raju, Antonio Torralba, Aude Oliva
  • BubbLeNet: Foveated Imaging for Visual Discovery [pdf] - Kevin Matzen, Noah Snavely
  • A Versatile Learning-Based 3D Temporal Tracker: Scalable, Robust, Online [pdf] - David Joseph Tan, Federico Tombari, Slobodan Ilic, Nassir Navab
  • Differential Recurrent Neural Networks for Action Recognition [pdf] - Vivek Veeriah, Naifan Zhuang, Guo-Jun Qi
Infinite Feature Selection [pdf]
Giorgio Roffo, Simone Melzi, Marco Cristani

Abstract: Filter-based feature selection has become crucial in many classification settings, especially object recognition, recently faced with feature learning strategies that originate thousands of cues. In this paper, we propose a feature selection method exploiting the convergence properties of power series of matrices, and introducing the concept of infinite feature selection (Inf-FS). Considering a selection of features as a path among feature distributions and letting these paths tend to an infinite number permits the investigation of the importance (relevance and redundancy) of a feature when injected into an arbitrary set of cues. Ranking the importance individuates candidate features, which turn out to be effective from a classification point of view, as proved by a thoroughly experimental section. The Inf-FS has been tested on thirteen diverse benchmarks, comparing against filters, embedded methods, and wrappers; in all the cases we achieve top performances, notably on the classification tasks of PASCAL VOC 2007-2012.
Similar papers:
  • Attribute-Graph: A Graph Based Approach to Image Ranking [pdf] - Nikita Prabhu, R. Venkatesh Babu
  • Live Repetition Counting [pdf] - Ofir Levy, Lior Wolf
  • Neural Activation Constellations: Unsupervised Part Model Discovery With Convolutional Networks [pdf] - Marcel Simon, Erik Rodner
  • Learning to Rank Based on Subsequences [pdf] - Basura Fernando, Efstratios Gavves, Damien Muselet, Tinne Tuytelaars
  • Structured Feature Selection [pdf] - Tian Gao, Ziheng Wang, Qiang Ji
Understanding Everyday Hands in Action From RGB-D Images [pdf]
Gregory Rogez, James S. Supancic III, Deva Ramanan

Abstract: We analyze functional manipulations of handheld objects, formalizing the problem as one of fine-grained grasp classification. To do so, we make use of a recently developed fine-grained taxonomy of human-object grasps. We introduce a large dataset of 12000 RGB-D images covering 71 everyday grasps in natural interactions. Our dataset is different from past work (typically addressed from a robotics perspective) in terms of its scale, diversity, and combination of RGB and depth data. From a computer-vision perspective, our dataset allows for exploration of contact and force prediction (crucial concepts in functional grasp analysis) from perceptual cues. We present extensive experimental results with state-of-the-art baselines, illustrating the role of segmentation, object context, and 3D-understanding in functional grasp analysis. We demonstrate a near 2X improvement over prior work and a naive deep baseline, while pointing out important directions for improvement.
Similar papers:
  • 3D Object Reconstruction From Hand-Object Interactions [pdf] - Dimitrios Tzionas, Juergen Gall
  • Training a Feedback Loop for Hand Pose Estimation [pdf] - Markus Oberweger, Paul Wohlhart, Vincent Lepetit
  • Lending A Hand: Detecting Hands and Recognizing Activities in Complex Egocentric Interactions [pdf] - Sven Bambach, Stefan Lee, David J. Crandall, Chen Yu
  • A Collaborative Filtering Approach to Real-Time Hand Pose Estimation [pdf] - Chiho Choi, Ayan Sinha, Joon Hee Choi, Sujin Jang, Karthik Ramani
  • Depth-Based Hand Pose Estimation: Data, Methods, and Challenges [pdf] - James S. Supancic III, Gregory Rogez, Yi Yang, Jamie Shotton, Deva Ramanan
From Emotions to Action Units With Hidden and Semi-Hidden-Task Learning [pdf]
Adria Ruiz, Joost Van de Weijer, Xavier Binefa

Abstract: Limited annotated training data is a challenging problem in Action Unit recognition. In this paper, we investigate how the use of large databases labelled according to the 6 universal facial expressions can increase the generalization ability of Action Unit classifiers. For this purpose, we propose a novel learning framework: Hidden-Task Learning. HTL aims to learn a set of Hidden-Tasks (Action Units) for which samples are not available but, in contrast, training data is easier to obtain from a set of related VisibleTasks (Facial Expressions). To that end, HTL is able to exploit prior knowledge about the relation between Hidden and Visible-Tasks. In our case, we base this prior knowledge on empirical psychological studies providing statistical correlations between Action Units and universal facial expressions. Additionally, we extend HTL to Semi-Hidden Task Learning (SHTL) assuming that Action Unit training samples are also provided. Performing exhaustive experiments over four different datasets, we show that HTL and SHTL improve the generalization ability of AU classifiers by training them with additional facial expression data. Additionally, we show that SHTL achieves competitive performance compared with state-of-the-art Transductive Learning approaches which face the problem of limited training data by using unlabelled test samples during training.
Similar papers:
  • FaceDirector: Continuous Control of Facial Performance in Video [pdf] - Charles Malleson, Jean-Charles Bazin, Oliver Wang, Derek Bradley, Thabo Beeler, Adrian Hilton, Alexander Sorkine-Hornung
  • Pairwise Conditional Random Forests for Facial Expression Recognition [pdf] - Arnaud Dapogny, Kevin Bailly, Severine Dubuisson
  • Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf] - Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic
  • Learning to Transfer: Transferring Latent Task Structures and Its Application to Person-Specific Facial Action Unit Detection [pdf] - Timur Almaev, Brais Martinez, Michel Valstar
  • Joint Fine-Tuning in Deep Neural Networks for Facial Expression Recognition [pdf] - Heechul Jung, Sihaeng Lee, Junho Yim, Sunjeong Park, Junmo Kim
Robust Statistical Face Frontalization [pdf]
Christos Sagonas, Yannis Panagakis, Stefanos Zafeiriou, Maja Pantic

Abstract: b) Optimization a) Input c) Output 1 Recently, it has been shown that excellent results can be achieved in both facial landmark localization and poseinvariant face recognition. These breakthroughs are attributed to the efforts of the community to manually annotate facial images in many different poses and to collect 3D facial data. In this paper, we propose a novel method for joint frontal view reconstruction and landmark localization using a small set of frontal images only. By observing that the frontal facial image is the one having the minimum rank of all different poses, an appropriate model which is able to jointly recover the frontalized version of the face as well as the facial landmarks is devised. To this end, a suitable optimization problem, involving the minimization of the nuclear norm and the matrix 1 norm is solved. The proposed method is assessed in frontal face reconstruction, face landmark localization, pose-invariant face recognition, and face verification in unconstrained conditions. The relevant experiments have been conducted on 8 databases. The experimental results demonstrate the effectiveness of the proposed method in comparison to the state-of-the-art methods for the target problems.
Similar papers:
  • Simultaneous Local Binary Feature Learning and Encoding for Face Recognition [pdf] - Jiwen Lu, Venice Erin Liong, Jie Zhou
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
  • Discriminative Pose-Free Descriptors for Face and Object Matching [pdf] - Soubhik Sanyal, Sivaram Prasad Mudunuri, Soma Biswas
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
Learning a Descriptor-Specific 3D Keypoint Detector [pdf]
Samuele Salti, Federico Tombari, Riccardo Spezialetti, Luigi Di Stefano

Abstract: Keypoint detection represents the first stage in the majority of modern computer vision pipelines based on automatically established correspondences between local descriptors. However, no standard solution has emerged yet in the case of 3D data such as point clouds or meshes, which exhibit high variability in level of detail and noise. More importantly, existing proposals for 3D keypoint detection rely on geometric saliency functions that attempt to maximize repeatability rather than distinctiveness of the selected regions, which may lead to sub-optimal performance of the overall pipeline. To overcome these shortcomings, we cast 3D keypoint detection as a binary classification between points whose support can be correctly matched by a predefined 3D descriptor or not, thereby learning a descriptorspecific detector that adapts seamlessly to different scenarios. Through experiments on several public datasets, we show that this novel approach to the design of a keypoint detector represents a flexible solution that, nonetheless, can provide state-of-the-art descriptor matching performance.
Similar papers:
  • An MRF-Poselets Model for Detecting Highly Articulated Humans [pdf] - Duc Thanh Nguyen, Minh-Khoi Tran, Sai-Kit Yeung
  • Predicting Good Features for Image Geo-Localization Using Per-Bundle VLAD [pdf] - Hyo Jin Kim, Enrique Dunn, Jan-Michael Frahm
  • Cluster-Based Point Set Saliency [pdf] - Flora Ponjou Tasse, Jiri Kosinka, Neil Dodgson
  • FASText: Efficient Unconstrained Scene Text Detector [pdf] - Michal Busta, Lukas Neumann, Jiri Matas
  • Multi-View Convolutional Neural Networks for 3D Shape Recognition [pdf] - Hang Su, Subhransu Maji, Evangelos Kalogerakis, Erik Learned-Miller
Naive Bayes Super-Resolution Forest [pdf]
Jordi Salvador, Eduardo Perez-Pellitero

Abstract: 32.5 NBSRF 32.4
Similar papers:
  • Convolutional Sparse Coding for Image Super-Resolution [pdf] - Shuhang Gu, Wangmeng Zuo, Qi Xie, Deyu Meng, Xiangchu Feng, Lei Zhang
  • External Patch Prior Guided Internal Clustering for Image Denoising [pdf] - Fei Chen, Lei Zhang, Huimin Yu
  • Learning Parametric Distributions for Image Super-Resolution: Where Patch Matching Meets Sparse Coding [pdf] - Yongbo Li, Weisheng Dong, Guangming Shi, Xuemei Xie
  • Deep Networks for Image Super-Resolution With Sparse Prior [pdf] - Zhaowen Wang, Ding Liu, Jianchao Yang, Wei Han, Thomas Huang
  • Deep Neural Decision Forests [pdf] - Peter Kontschieder, Madalina Fiterau, Antonio Criminisi, Samuel Rota Bulo
Discriminative Pose-Free Descriptors for Face and Object Matching [pdf]
Soubhik Sanyal, Sivaram Prasad Mudunuri, Soma Biswas

Abstract: Pose invariant matching is a very important and challenging problem with various applications like recognizing faces in uncontrolled scenarios, matching objects taken from different view points, etc. In this paper, we propose a discriminative pose-free descriptor (DPFD) which can be used to match faces/objects across pose variations. Training examples at very few representative poses are used to generate virtual intermediate pose subspaces. An image or image region is then represented by a feature set obtained by projecting it on all these subspaces and a discriminative transform is applied on this feature set to make it suitable for classification tasks. Finally, this discriminative feature set is represented by a single feature vector, termed as DPFD. The DPFD of images taken from different viewpoints can be directly compared for matching. Extensive experiments on recognizing faces across pose, pose and resolution on the Multi-PIE and Surveillance Cameras Face datasets and comparisons with state-of-the-art approaches show the effectiveness of the proposed approach. Experiments on matching general objects across viewpoints show the generalizability of the proposed approach beyond faces.
Similar papers:
  • Learning Where to Position Parts in 3D [pdf] - Marco Pedersoli, Tinne Tuytelaars
  • Pose Induction for Novel Object Categories [pdf] - Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
  • Robust Statistical Face Frontalization [pdf] - Christos Sagonas, Yannis Panagakis, Stefanos Zafeiriou, Maja Pantic
Hyperpoints and Fine Vocabularies for Large-Scale Location Recognition [pdf]
Torsten Sattler, Michal Havlena, Filip Radenovic, Konrad Schindler, Marc Pollefeys

Abstract: Structure-based localization is the task of finding the absolute pose of a given query image w.r.t. a pre-computed 3D model. While this is almost trivial at small scale, special care must be taken as the size of the 3D model grows, because straight-forward descriptor matching becomes ineffective due to the large memory footprint of the model, as well as the strictness of the ratio test in 3D. Recently, several authors have tried to overcome these problems, either by a smart compression of the 3D model or by clever sampling strategies for geometric verification. Here we explore an orthogonal strategy, which uses all the 3D points and standard sampling, but performs feature matching implicitly, by quantization into a fine vocabulary. We show that although this matching is ambiguous and gives rise to 3D hyperpoints when matching each 2D query feature in isolation, a simple voting strategy, which enforces the fact that the selected 3D points shall be co-visible, can reliably find a locally unique 2D-3D point assignment. Experiments on two large-scale datasets demonstrate that our method achieves state-of-the-art performance, while the memory footprint is greatly reduced, since only visual word labels but no 3D point descriptors need to be stored.
Similar papers:
  • Scalable Person Re-Identification: A Benchmark [pdf] - Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian
  • Multi-Image Matching via Fast Alternating Minimization [pdf] - Xiaowei Zhou, Menglong Zhu, Kostas Daniilidis
  • Predicting Good Features for Image Geo-Localization Using Per-Bundle VLAD [pdf] - Hyo Jin Kim, Enrique Dunn, Jan-Michael Frahm
  • Localize Me Anywhere, Anytime: A Multi-Task Point-Retrieval Approach [pdf] - Guoyu Lu, Yan Yan, Li Ren, Jingkuan Song, Nicu Sebe, Chandra Kambhamettu
  • Camera Pose Voting for Large-Scale Image-Based Localization [pdf] - Bernhard Zeisl, Torsten Sattler, Marc Pollefeys
Unsupervised Generation of a Viewpoint Annotated Car Dataset From Videos [pdf]
Nima Sedaghat, Thomas Brox

Abstract: Object recognition approaches have recently been extended to yield, aside of the object class output, also viewpoint or pose. Training such approaches typically requires additional viewpoint or keypoint annotation in the training data or, alternatively, synthetic CAD models. In this paper, we present an approach that creates a dataset of images annotated with bounding boxes and viewpoint labels in a fully automated manner from videos. We assume that the scene is static in order to reconstruct 3D surfaces via structure from motion. We automatically detect when the reconstruction fails and normalize for the viewpoint of the 3D models by aligning the reconstructed point clouds. Exemplarily for cars we show that we can expand a large dataset of annotated single images and obtain improved performance when training a viewpoint regressor on this joined dataset.
Similar papers:
  • Single Image Pop-Up From Discriminatively Learned Parts [pdf] - Menglong Zhu, Xiaowei Zhou, Kostas Daniilidis
  • Continuous Pose Estimation With a Spatial Ensemble of Fisher Regressors [pdf] - Michele Fenzi, Laura Leal-Taixe, Jorn Ostermann, Tinne Tuytelaars
  • Scene-Domain Active Part Models for Object Representation [pdf] - Zhou Ren, Chaohui Wang, Alan L. Yuille
  • Real-Time Pose Estimation Piggybacked on Object Detection [pdf] - Roman Juranek, Adam Herout, Marketa Dubska, Pavel Zemcik
  • Render for CNN: Viewpoint Estimation in Images Using CNNs Trained With Rendered 3D Model Views [pdf] - Hao Su, Charles R. Qi, Yangyan Li, Leonidas J. Guibas
Unsupervised Semantic Parsing of Video Collections [pdf]
Ozan Sener, Amir R. Zamir, Silvio Savarese, Ashutosh Saxena

Abstract: Human communication typically has an underlying structure. This is reflected in the fact that in many user generated videos, a starting point, ending, and certain objective steps between these two can be identified. In this paper, we propose a method for parsing a video into such semantic steps in an unsupervised way. The proposed method is capable of providing a semantic storyline of the video composed of its objective steps. We accomplish this using both visual and language cues in a joint generative model. The proposed method can also provide a textual description for each of the identified semantic steps. We evaluate this method on a large number of complex YouTube videos and show results of unprecedented quality for this intricate and impactful problem.
Similar papers:
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
  • Context Aware Active Learning of Activity Recognition Models [pdf] - Mahmudul Hasan, Amit K. Roy-Chowdhury
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
  • Temporal Perception and Prediction in Ego-Centric Video [pdf] - Yipin Zhou, Tamara L. Berg
  • Activity Auto-Completion: Predicting Human Activities From Partial Videos [pdf] - Zhen Xu, Laiyun Qing, Jun Miao
Robust Principal Component Analysis on Graphs [pdf]
Nauman Shahid, Vassilis Kalofolias, Xavier Bresson, Michael Bronstein, Pierre Vandergheynst

Abstract: Principal Component Analysis (PCA) is the most widely used tool for linear dimensionality reduction and clustering. Still it is highly sensitive to outliers and does not scale well with respect to the number of data samples. Robust PCA solves the first issue with a sparse penalty term. The second issue can be handled with the matrix factorization model, which is however non-convex. Besides, PCA based clustering can also be enhanced by using a graph of data similarity. In this article, we introduce a new model called ‘Robust PCA on Graphs’ which incorporates spectral graph regularization into the Robust PCA framework. Our proposed model benefits from 1) the robustness of principal components to occlusions and missing values, 2) enhanced lowrank recovery, 3) improved clustering property due to the graph smoothness assumption on the low-rank matrix, and 4) convexity of the resulting optimization problem. Extensive experiments on 8 benchmark, 3 video and 2 artificial datasets with corruptions clearly reveal that our model outperforms 10 other state-of-the-art models in its clustering and low-rank recovery tasks.
Similar papers:
  • Mining And-Or Graphs for Graph Matching and Object Discovery [pdf] - Quanshi Zhang, Ying Nian Wu, Song-Chun Zhu
  • Shape Interaction Matrix Revisited and Robustified: Efficient Subspace Clustering With Corrupted and Incomplete Data [pdf] - Pan Ji, Mathieu Salzmann, Hongdong Li
  • Temporal Subspace Clustering for Human Motion Segmentation [pdf] - Sheng Li, Kang Li, Yun Fu
  • Shell PCA: Statistical Shape Modelling in Shell Space [pdf] - Chao Zhang, Behrend Heeren, Martin Rumpf, William A. P. Smith
  • A Matrix Decomposition Perspective to Multiple Graph Matching [pdf] - Junchi Yan, Hongteng Xu, Hongyuan Zha, Xiaokang Yang, Huanxi Liu, Stephen Chu
Classical Scaling Revisited [pdf]
Gil Shamai, Yonathan Aflalo, Michael Zibulevsky, Ron Kimmel

Abstract: Multidimensional-scaling (MDS) is an information analysis tool. It involves the evaluation of distances between data points, which is a quadratic space-time problem. Then, MDS procedures find an embedding of the points in a low dimensional Euclidean (flat) domain, optimizing for the similarity of inter-points distances. We present an efficient solver for Classical Scaling (a specific MDS model) by extending the distances measured from a subset of the points to the rest, while exploiting the smoothness property of the distance functions. The smoothness is measured by the L2 norm of the Laplace-Beltrami operator applied to the unknown distance function. The Laplace Beltrami reflects the local differential relations between points, and can be computed in linear time. Classical-scaling is thereby reformulated into a quasi-linear space-time complexities procedure.
Similar papers:
  • A Projection Free Method for Generalized Eigenvalue Problem With a Nonsmooth Regularizer [pdf] - Seong Jae Hwang, Maxwell D. Collins, Sathya N. Ravi, Vamsi K. Ithapu, Nagesh Adluru, Sterling C. Johnson, Vikas Singh
  • Beyond Gauss: Image-Set Matching on the Riemannian Manifold of PDFs [pdf] - Mehrtash Harandi, Mathieu Salzmann, Mahsa Baktashmotlagh
  • Shell PCA: Statistical Shape Modelling in Shell Space [pdf] - Chao Zhang, Behrend Heeren, Martin Rumpf, William A. P. Smith
  • Recursive Frechet Mean Computation on the Grassmannian and its Applications to Computer Vision [pdf] - Rudrasis Chakraborty, Baba C. Vemuri
  • Piecewise Flat Embedding for Image Segmentation [pdf] - Yizhou Yu, Chaowei Fang, Zicheng Liao
Scalable Nonlinear Embeddings for Semantic Category-Based Image Retrieval [pdf]
Gaurav Sharma, Bernt Schiele

Abstract: We propose a novel algorithm for the task of supervised discriminative distance learning by nonlinearly embedding vectors into a low dimensional Euclidean space. We work in the challenging setting where supervision is with constraints on similar and dissimilar pairs while training. The proposed method is derived by an approximate kernelization of a linear Mahalanobis-like distance metric learning algorithm and can also be seen as a kernel neural network. The number of model parameters and test time evaluation complexity of the proposed method are O(dD) where D is the dimensionality of the input features and d is the dimension of the projection space–this is in contrast to the usual kernelization methods as, unlike them, the complexity does not scale linearly with the number of training examples. We propose a stochastic gradient based learning algorithm which makes the method scalable (w.r.t. the number of training examples), while being nonlinear. We train the method with up to half a million training pairs of 4096 dimensional CNN features. We give empirical comparisons with relevant baselines on seven challenging datasets for the task of low dimensional semantic category based image retrieval.
Similar papers:
  • BodyPrint: Pose Invariant 3D Shape Matching of Human Bodies [pdf] - Jiangping Wang, Kai Ma, Vivek Kumar Singh, Thomas Huang, Terrence Chen
  • A Supervised Low-Rank Method for Learning Invariant Subspaces [pdf] - Farzad Siyahjani, Ranya Almohsen, Sinan Sabri, Gianfranco Doretto
  • Contractive Rectifier Networks for Nonlinear Maximum Margin Classification [pdf] - Senjian An, Munawar Hayat, Salman H. Khan, Mohammed Bennamoun, Farid Boussaid, Ferdous Sohel
  • Geometry-Aware Deep Transform [pdf] - Jiaji Huang, Qiang Qiu, Robert Calderbank, Guillermo Sapiro
  • Efficient PSD Constrained Asymmetric Metric Learning for Person Re-Identification [pdf] - Shengcai Liao, Stan Z. Li
Learning Binary Codes for Maximum Inner Product Search [pdf]
Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen

Abstract: Binary coding or hashing techniques are recognized to accomplish efficient near neighbor search, and have thus attracted broad interests in the recent vision and learning studies. However, such studies have rarely been dedicated to Maximum Inner Product Search (MIPS), which plays a critical role in various vision applications. In this paper, we investigate learning binary codes to exclusively handle the MIPS problem. Inspired by the latest advance in asymmetric hashing schemes, we propose an asymmetric binary code learning framework based on inner product fitting. Specifically, two sets of coding functions are learned such that the inner products between their generated binary codes can reveal the inner products between original data vectors. We also propose an alternative simpler objective which maximizes the correlations between the inner products of the produced binary codes and raw data vectors. In both objectives, the binary codes and coding functions are simultaneously learned without continuous relaxations, which is the key to achieving high-quality binary codes. We evaluate the proposed method, dubbed Asymmetric Innerproduct Binary Coding (AIBC), relying on the two objectives on several large-scale image datasets. Both of them are superior to the state-of-the-art binary coding and hashing methods in performing MIPS tasks.
Similar papers:
  • An NMF Perspective on Binary Hashing [pdf] - Lopamudra Mukherjee, Sathya N. Ravi, Vamsi K. Ithapu, Tyler Holmes, Vikas Singh
  • Adaptive Hashing for Fast Similarity Search [pdf] - Fatih Cakir, Stan Sclaroff
  • kNN Hashing With Factorized Neighborhood Representation [pdf] - Kun Ding, Chunlei Huo, Bin Fan, Chunhong Pan
  • Top Rank Supervised Binary Coding for Visual Search [pdf] - Dongjin Song, Wei Liu, Rongrong Ji, David A. Meyer, John R. Smith
  • Alternating Co-Quantization for Cross-Modal Hashing [pdf] - Go Irie, Hiroyuki Arai, Yukinobu Taniguchi
Mutual-Structure for Joint Filtering [pdf]
Xiaoyong Shen, Chao Zhou, Li Xu, Jiaya Jia

Abstract: Previous joint/guided filters directly transfer the structural information in the reference image to the target one. In this paper, we first analyze its major drawback that is, there may be completely different edges in the two images. Simply passing all patterns to the target could introduce significant errors. To address this issue, we propose the concept of mutual-structure, which refers to the structural information that is contained in both images and thus can be safely enhanced by joint filtering, and an untraditional objective function that can be efficiently optimized to yield mutual structure. Our method results in necessary and important edge preserving, which greatly benefits depth completion, optical flow estimation, image enhancement, stereo matching, to name a few. (a) Input (b) Reference (c) Bilateral Filter (d) Guided Filter (e) Ours (f) Ground Truth Figure 1. Joint image filtering on a structure-inconsistent image pair. (a) and (b) are the target and reference images respectively. (c) and (d) are the results of bilateral and guided image filtering respectively, which transfer color-image edges to depth. (e) is our result that does not contain the texture patterns and unwanted edges from (b), and (f) is the ground truth.
Similar papers:
  • Learning Nonlinear Spectral Filters for Color Image Reconstruction [pdf] - Michael Moeller, Julia Diebold, Guy Gilboa, Daniel Cremers
  • Convolutional Sparse Coding for Image Super-Resolution [pdf] - Shuhang Gu, Wangmeng Zuo, Qi Xie, Deyu Meng, Xiangchu Feng, Lei Zhang
  • Deep Colorization [pdf] - Zezhou Cheng, Qingxiong Yang, Bin Sheng
  • Fully Connected Guided Image Filtering [pdf] - Longquan Dai, Mengke Yuan, Feihu Zhang, Xiaopeng Zhang
  • Segment Graph Based Image Filtering: Fast Structure-Preserving Smoothing [pdf] - Feihu Zhang, Longquan Dai, Shiming Xiang, Xiaopeng Zhang
Person Re-Identification With Correspondence Structure Learning [pdf]
Yang Shen, Weiyao Lin, Junchi Yan, Mingliang Xu, Jianxin Wu, Jingdong Wang

Abstract: This paper addresses the problem of handling spatial misalignments due to camera-view changes or human-pose variations in person re-identification. We first introduce a boosting-based approach to learn a correspondence structure which indicates the patch-wise matching probabilities between images from a target camera pair. The learned correspondence structure can not only capture the spatial correspondence pattern between cameras but also handle the viewpoint or human-pose variation in individual images. We further introduce a global-based matching process. It integrates a global matching constraint over the learned correspondence structure to exclude cross-view misalignments during the image patch matching process, hence achieving a more reliable matching score between images. Experimental results on various datasets demonstrate the effectiveness of our approach.
Similar papers:
  • Scalable Person Re-Identification: A Benchmark [pdf] - Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian
  • Dense Semantic Correspondence Where Every Pixel is a Classifier [pdf] - Hilton Bristow, Jack Valmadre, Simon Lucey
  • Person Re-Identification With Discriminatively Trained Viewpoint Invariant Dictionaries [pdf] - Srikrishna Karanam, Yang Li, Richard J. Radke
  • Multi-Scale Learning for Low-Resolution Person Re-Identification [pdf] - Xiang Li, Wei-Shi Zheng, Xiaojuan Wang, Tao Xiang, Shaogang Gong
  • Partial Person Re-Identification [pdf] - Wei-Shi Zheng, Xiang Li, Tao Xiang, Shengcai Liao, Jianhuang Lai, Shaogang Gong
Personalized Age Progression With Aging Dictionary [pdf]
Xiangbo Shu, Jinhui Tang, Hanjiang Lai, Luoqi Liu, Shuicheng Yan

Abstract: In this paper, we aim to automatically render aging faces in a personalized way. Basically, a set of age-group specific dictionaries are learned, where the dictionary bases corresponding to the same index yet from different dictionaries form a particular aging process pattern cross different age groups, and a linear combination of these patterns expresses a particular personalized aging process. Moreover, two factors are taken into consideration in the dictionary learning process. First, beyond the aging dictionaries, each subject may have extra personalized facial characteristics, e.g. mole, which are invariant in the aging process. Second, it is challenging or even impossible to collect faces of all age groups for a particular subject, yet much easier and more practical to get face pairs from neighboring age groups. Thus a personality-aware coupled reconstruction loss is utilized to learn the dictionaries based on face pairs from neighboring age groups. Extensive experiments well demonstrate the advantages of our proposed solution over other state-of-the-arts in term of personalized aging progression, as well as the performance gain for cross-age face verification by synthesizing aging faces. This work was performed when X. Shu was visiting National University of Singapore. Corresponding author.
Similar papers:
  • Learning to Predict Saliency on Face Images [pdf] - Mai Xu, Yun Ren, Zulin Wang
  • Deep Learning Face Attributes in the Wild [pdf] - Ziwei Liu, Ping Luo, Xiaogang Wang, Xiaoou Tang
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • Simultaneous Local Binary Feature Learning and Encoding for Face Recognition [pdf] - Jiwen Lu, Venice Erin Liong, Jie Zhou
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
Discriminative Learning of Deep Convolutional Feature Point Descriptors [pdf]
Edgar Simo-Serra, Eduard Trulls, Luis Ferraz, Iasonas Kokkinos, Pascal Fua, Francesc Moreno-Noguer

Abstract: Deep learning has revolutionalized image-level tasks such as classification, but patch-level tasks, such as correspondence, still rely on hand-crafted features, e.g. SIFT. In this paper we use Convolutional Neural Networks (CNNs) to learn discriminant patch representations and in particular train a Siamese network with pairs of (non-)corresponding patches. We deal with the large number of potential pairs with the combination of a stochastic sampling of the training set and an aggressive mining strategy biased towards patches that are hard to classify. By using the L2 distance during both training and testing we develop 128-D descriptors whose euclidean distances reflect patch similarity, and which can be used as a drop-in replacement for any task involving SIFT. We demonstrate consistent performance gains over the state of the art, and generalize well against scaling and rotation, perspective transformation, non-rigid deformation, and illumination changes. Our descriptors are efficient to compute and amenable to modern GPUs, and are publicly available. (a) Data (b) All pairs (c) Hard pairs Figure 1: To train models with Siamese networks, we need pairs of corresponding and non-corresponding samples. (a) We use t-SNE [30] to display 100 64 64 patches of 12 3D points from different images (see Fig. 3 for examples). Corresponding patches are drawn with the same color. (b) We single out the red-circled patch, belonging to the blue point cloud, and consider all of its pote
Similar papers:
  • Unsupervised Learning of Visual Representations Using Videos [pdf] - Xiaolong Wang, Abhinav Gupta
  • Unsupervised Visual Representation Learning by Context Prediction [pdf] - Carl Doersch, Abhinav Gupta, Alexei A. Efros
  • Aggregating Local Deep Features for Image Retrieval [pdf] - Artem Babenko, Victor Lempitsky
  • Local Convolutional Features With Unsupervised Training for Image Retrieval [pdf] - Mattis Paulin, Matthijs Douze, Zaid Harchaoui, Julien Mairal, Florent Perronin, Cordelia Schmid
  • RIDE: Reversal Invariant Descriptor Enhancement [pdf] - Lingxi Xie, Jingdong Wang, Weiyao Lin, Bo Zhang, Qi Tian
Neural Activation Constellations: Unsupervised Part Model Discovery With Convolutional Networks [pdf]
Marcel Simon, Erik Rodner

Abstract: Part models of object categories are essential for challenging recognition tasks, where differences in categories are subtle and only reflected in appearances of small parts of the object. We present an approach that is able to learn part models in a completely unsupervised manner, without part annotations and even without given bounding boxes during learning. The key idea is to find constellations of neural activation patterns computed using convolutional neural networks. In our experiments, we outperform existing approaches for fine-grained recognition on the CUB200-2011, Oxford PETS, and Oxford Flowers dataset in case no part or bounding box annotations are available and achieve state-of-the-art performance for the Stanford Dog dataset. We also show the benefits of neural constellation models as a data augmentation technique for fine-tuning. Furthermore, our paper unites the areas of generic and fine-grained classification, since our approach is suitable for both scenarios.
Similar papers:
  • Cascaded Sparse Spatial Bins for Efficient and Effective Generic Object Detection [pdf] - David Novotny, Jiri Matas
  • Augmenting Strong Supervision Using Web Data for Fine-Grained Categorization [pdf] - Zhe Xu, Shaoli Huang, Ya Zhang, Dacheng Tao
  • Box Aggregation for Proposal Decimation: Last Mile of Object Detection [pdf] - Shu Liu, Cewu Lu, Jiaya Jia
  • Learning Where to Position Parts in 3D [pdf] - Marco Pedersoli, Tinne Tuytelaars
  • Actions and Attributes From Wholes and Parts [pdf] - Georgia Gkioxari, Ross Girshick, Jitendra Malik
Selecting Relevant Web Trained Concepts for Automated Event Retrieval [pdf]
Bharat Singh, Xintong Han, Zhe Wu, Vlad I. Morariu, Larry S. Davis

Abstract: Complex event retrieval is a challenging research problem, especially when no training videos are available. An alternative to collecting training videos is to train a large semantic concept bank a priori. Given a text description of an event, event retrieval is performed by selecting concepts linguistically related to the event description and fusing the concept responses on unseen videos. However, defining an exhaustive concept lexicon and pre-training it requires vast computational resources. Therefore, recent approaches automate concept discovery and training by leveraging large amounts of weakly annotated web data. Compact visually salient concepts are automatically obtained by the use of concept pairs or, more generally, n-grams. However, not all visually salient n-grams are necessarily useful for an event querysome combinations of concepts may be visually compact but irrelevantand this drastically affects performance. We propose an event retrieval algorithm that constructs pairs of automatically discovered concepts and then prunes those concepts that are unlikely to be helpful for retrieval. Pruning depends both on the query and on the specific video instance being evaluated. Our approach also addresses calibration and domain adaptation issues that arise when applying concept detectors to unseen videos. We demonstrate large improvements over other vision based systems on the TRECVID MED 13 dataset.
Similar papers:
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
  • Towards Computational Baby Learning: A Weakly-Supervised Approach for Object Detection [pdf] - Xiaodan Liang, Si Liu, Yunchao Wei, Luoqi Liu, Liang Lin, Shuicheng Yan
  • LEWIS: Latent Embeddings for Word Images and their Semantics [pdf] - Albert Gordo, Jon Almazan, Naila Murray, Florent Perronin
  • Learning Like a Child: Fast Novel Visual Concept Learning From Sentence Descriptions of Images [pdf] - Junhua Mao, Xu Wei, Yi Yang, Jiang Wang, Zhiheng Huang, Alan L. Yuille
  • Automatic Concept Discovery From Parallel Text and Visual Corpora [pdf] - Chen Sun, Chuang Gan, Ram Nevatia
Projection Onto the Manifold of Elongated Structures for Accurate Extraction [pdf]
Amos Sironi, Vincent Lepetit, Pascal Fua

Abstract: Detection of elongated structures in 2D images and 3D image stacks is a critical prerequisite in many applications and Machine Learning-based approaches have recently been shown to deliver superior performance. However, these methods essentially classify individual locations and do not explicitly model the strong relationship that exists between neighboring ones. As a result, isolated erroneous responses, discontinuities, and topological errors are present in the resulting score maps. We solve this problem by projecting patches of the score map to their nearest neighbors in a set of ground truth training patches. Our algorithm induces global spatial consistency on the classifier score map and returns results that are provably geometrically consistent. We apply our algorithm to challenging datasets in four different domains and show that it compares favorably to state-of-the-art methods.
Similar papers:
  • External Patch Prior Guided Internal Clustering for Image Denoising [pdf] - Fei Chen, Lei Zhang, Huimin Yu
  • Unsupervised Visual Representation Learning by Context Prediction [pdf] - Carl Doersch, Abhinav Gupta, Alexei A. Efros
  • Harvesting Discriminative Meta Objects With Deep CNN Features for Scene Classification [pdf] - Ruobing Wu, Baoyuan Wang, Wenping Wang, Yizhou Yu
  • One Shot Learning via Compositions of Meaningful Patches [pdf] - Alex Wong, Alan L. Yuille
  • High-for-Low and Low-for-High: Efficient Boundary Detection From Deep Object Features and its Applications to High-Level Vision [pdf] - Gedas Bertasius, Jianbo Shi, Lorenzo Torresani
A Supervised Low-Rank Method for Learning Invariant Subspaces [pdf]
Farzad Siyahjani, Ranya Almohsen, Sinan Sabri, Gianfranco Doretto

Abstract: Sparse representation and low-rank matrix decomposition approaches have been successfully applied to several computer vision problems. They build a generative representation of the data, which often requires complex training as well as testing to be robust against data variations induced by nuisance factors. We introduce the invariant components, a discriminative representation invariant to nuisance factors, because it spans subspaces orthogonal to the space where nuisance factors are defined. This allows developing a framework based on geometry that ensures a uniform inter-class separation, and a very efficient and robust classification based on simple nearest neighbor. In addition, we show how the approach is equivalent to a local metric learning, where the local metrics (one for each class) are learned jointly, rather than independently, thus avoiding the risk of overfitting without the need for additional regularization. We evaluated the approach for face recognition with highly corrupted training and testing data, obtaining very promising results.
Similar papers:
  • Geometry-Aware Deep Transform [pdf] - Jiaji Huang, Qiang Qiu, Robert Calderbank, Guillermo Sapiro
  • Personalized Age Progression With Aging Dictionary [pdf] - Xiangbo Shu, Jinhui Tang, Hanjiang Lai, Luoqi Liu, Shuicheng Yan
  • Scalable Nonlinear Embeddings for Semantic Category-Based Image Retrieval [pdf] - Gaurav Sharma, Bernt Schiele
  • Efficient PSD Constrained Asymmetric Metric Learning for Person Re-Identification [pdf] - Shengcai Liao, Stan Z. Li
  • Simultaneous Local Binary Feature Learning and Encoding for Face Recognition [pdf] - Jiwen Lu, Venice Erin Liong, Jie Zhou
Face Flow [pdf]
Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou

Abstract: In this paper, we propose a method for the robust and efficient computation of multi-frame optical flow in an expressive sequence of facial images. We formulate a novel energy minimisation problem for establishing dense correspondences between a neutral template and every frame of a sequence. We exploit the highly correlated nature of human expressions by representing dense facial motion using a deformation basis. Furthermore, we exploit the even higher correlation between deformations in a given input sequence by imposing a low-rank prior on the coefficients of the deformation basis, yielding temporally consistent optical flow. Our proposed model-based formulation, in conjunction with the inverse compositional strategy and lowrank matrix optimisation that we adopt, leads to a highly efficient algorithm for calculating facial flow. As experimental evaluation, we show quantitative experiments on a challenging novel benchmark of face sequences, with dense ground truth optical flow provided by motion capture data. We also provide qualitative results on a real sequence displaying fast motion and occlusions. Extensive quantitative and qualitative comparisons demonstrate that the proposed method outperforms state-of-the-art optical flow and dense non-rigid registration techniques, whilst running an order of magnitude faster.
Similar papers:
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • FlowNet: Learning Optical Flow With Convolutional Networks [pdf] - Alexey Dosovitskiy, Philipp Fischer, Eddy Ilg, Philip Hausser, Caner Hazirbas, Vladimir Golkov, Patrick van der Smagt, Daniel Cremers, Thomas Brox
  • Large Displacement 3D Scene Flow With Occlusion Reasoning [pdf] - Andrei Zanfir, Cristian Sminchisescu
  • Oriented Light-Field Windows for Scene Flow [pdf] - Pratul P. Srinivasan, Michael W. Tao, Ren Ng, Ravi Ramamoorthi
  • Flow Fields: Dense Correspondence Fields for Highly Accurate Large Displacement Optical Flow Estimation [pdf] - Christian Bailer, Bertram Taetz, Didier Stricker
Learning to Divide and Conquer for Online Multi-Target Tracking [pdf]
Francesco Solera, Simone Calderara, Rita Cucchiara

Abstract: Online Multiple Target Tracking (MTT) is often addressed within the tracking-by-detection paradigm. Detections are previously extracted independently in each frame and then objects trajectories are built by maximizing specifically designed coherence functions. Nevertheless, ambiguities arise in presence of occlusions or detection errors. In this paper we claim that the ambiguities in tracking could be solved by a selective use of the features, by working with more reliable features if possible and exploiting a deeper representation of the target only if necessary. To this end, we propose an online divide and conquer tracker for static camera scenes, which partitions the assignment problem in local subproblems and solves them by selectively choosing and combining the best features. The complete framework is cast as a structural learning task that unifies these phases and learns tracker parameters from examples. Experiments on two different datasets highlights a significant improvement of tracking performances (MOTA +10%) over the state of the art.
Similar papers:
  • FollowMe: Efficient Online Min-Cost Flow Tracking With Bounded Memory and Computation [pdf] - Philip Lenz, Andreas Geiger, Raquel Urtasun
  • Multiple Hypothesis Tracking Revisited [pdf] - Chanho Kim, Fuxin Li, Arridhana Ciptadi, James M. Rehg
  • Joint Probabilistic Data Association Revisited [pdf] - Seyed Hamid Rezatofighi, Anton Milan, Zhen Zhang, Qinfeng Shi, Anthony Dick, Ian Reid
  • Learning to Track: Online Multi-Object Tracking by Decision Making [pdf] - Yu Xiang, Alexandre Alahi, Silvio Savarese
  • Near-Online Multi-Target Tracking With Aggregated Local Flow Descriptor [pdf] - Wongun Choi
Tracking-by-Segmentation With Online Gradient Boosting Decision Tree [pdf]
Jeany Son, Ilchae Jung, Kayoung Park, Bohyung Han

Abstract: We propose an online tracking algorithm that adaptively models target appearances based on an online gradient boosting decision tree. Our algorithm is particularly useful for non-rigid and/or articulated objects since it handles various deformations of the target effectively by integrating a classifier operating on individual patches and provides segmentation masks of the target as final results. The posterior of the target state is propagated over time by particle filtering, where the likelihood is computed based mainly on a patch-level confidence map associated with a latent target state corresponding to each sample. Once tracking is completed in each frame, our gradient boosting decision tree is updated to adapt new data in a recursive manner. For effective evaluation of segmentation-based tracking algorithms, we construct a new ground-truth that contains pixellevel annotation of segmentation mask. We evaluate the performance of our tracking algorithm based on the measures for segmentation masks, where our algorithm illustrates outstanding accuracy compared to the state-of-theart segmentation-based tracking methods.
Similar papers:
  • Near-Online Multi-Target Tracking With Aggregated Local Flow Descriptor [pdf] - Wongun Choi
  • Understanding and Diagnosing Visual Tracking Systems [pdf] - Naiyan Wang, Jianping Shi, Dit-Yan Yeung, Jiaya Jia
  • Visual Tracking With Fully Convolutional Networks [pdf] - Lijun Wang, Wanli Ouyang, Xiaogang Wang, Huchuan Lu
  • SOWP: Spatially Ordered and Weighted Patch Descriptor for Visual Tracking [pdf] - Han-Ul Kim, Dae-Youn Lee, Jae-Young Sim, Chang-Su Kim
  • Learning to Track: Online Multi-Object Tracking by Decision Making [pdf] - Yu Xiang, Alexandre Alahi, Silvio Savarese
Similarity Gaussian Process Latent Variable Model for Multi-Modal Data Analysis [pdf]
Guoli Song, Shuhui Wang, Qingming Huang, Qi Tian

Abstract: Data from real applications involve multiple modalities representing content with the same semantics and deliver rich information from complementary aspects. However, relations among heterogeneous modalities are simply treated as observation-to-t by existing work, and the parameterized cross-modal mapping functions lack flexibility in directly adapting to the content divergence and semantic complicacy of multi-modal data. In this paper, we build our work based on Gaussian process latent variable model (GPLVM) to learn the non-linear non-parametric mapping functions and transform heterogeneous data into a shared latent space. We propose multi-modal Similarity Gaussian Process latent variable model (m-SimGP), which learns the nonlinear mapping functions between the intra-modal similarities and latent representation. We further propose multimodal regularized similarity GPLVM (m-RSimGP) by encouraging similar/dissimilar points to be similar/dissimilar in the output space. The overall objective functions are solved by simple and scalable gradient decent techniques. The proposed models are robust to content divergence and high-dimensionality in multi-modal representation. They can be applied to various tasks to discover the non-linear correlations and obtain the comparable low-dimensional representation for heterogeneous modalities. On two widely used real-world datasets, we outperform previous approaches for cross-modal content retrieval and cross-modal classication.
Similar papers:
  • Entropy-Based Latent Structured Output Prediction [pdf] - Diane Bouchacourt, Sebastian Nowozin, M. Pawan Kumar
  • MANTRA: Minimum Maximum Latent Structural SVM for Image Classification and Ranking [pdf] - Thibaut Durand, Nicolas Thome, Matthieu Cord
  • A Nonparametric Bayesian Approach Toward Stacked Convolutional Independent Component Analysis [pdf] - Sotirios P. Chatzis, Dimitrios Kosmopoulos
  • Multi-Conditional Latent Variable Model for Joint Facial Action Unit Detection [pdf] - Stefanos Eleftheriadis, Ognjen Rudovic, Maja Pantic
  • Cutting Edge: Soft Correspondences in Multimodal Scene Parsing [pdf] - Sarah Taghavi Namin, Mohammad Najafi, Mathieu Salzmann, Lars Petersson
Top Rank Supervised Binary Coding for Visual Search [pdf]
Dongjin Song, Wei Liu, Rongrong Ji, David A. Meyer, John R. Smith

Abstract: In recent years, binary coding techniques are becoming increasingly popular because of their high efficiency in handling large-scale computer vision applications. It has been demonstrated that supervised binary coding techniques that leverage supervised information can significantly enhance the coding quality, and hence greatly benefit visual search tasks. Typically, a modern binary coding method seeks to learn a group of coding functions which compress data samples into binary codes. However, few methods pursued the coding functions such that the precision at the top of a ranking list according to Hamming distances of the generated binary codes is optimized. In this paper, we propose a novel supervised binary coding approach, namely Top Rank Supervised Binary Coding (Top-RSBC), which explicitly focuses on optimizing the precision of top positions in a Hamming-distance ranking list towards preserving the supervision information. The core idea is to train the disciplined coding functions, by which the mistakes at the top of a Hamming-distance ranking list are penalized more than those at the bottom. To solve such coding functions, we relax the original discrete optimization objective with a continuous surrogate, and derive a stochastic gradient descent to optimize the surrogate objective. To further reduce the training time cost, we also design an online learning algorithm to optimize the surrogate objective more efficiently. Empirical studies based upon three benchmark image
Similar papers:
  • Projection Bank: From High-Dimensional Data to Medium-Length Binary Codes [pdf] - Li Liu, Mengyang Yu, Ling Shao
  • kNN Hashing With Factorized Neighborhood Representation [pdf] - Kun Ding, Chunlei Huo, Bin Fan, Chunhong Pan
  • Alternating Co-Quantization for Cross-Modal Hashing [pdf] - Go Irie, Hiroyuki Arai, Yukinobu Taniguchi
  • An NMF Perspective on Binary Hashing [pdf] - Lopamudra Mukherjee, Sathya N. Ravi, Vamsi K. Ithapu, Tyler Holmes, Vikas Singh
  • Learning Binary Codes for Maximum Inner Product Search [pdf] - Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen
Action Localization in Videos Through Context Walk [pdf]
Khurram Soomro, Haroon Idrees, Mubarak Shah

Abstract: This paper presents an efficient approach for localizing actions by learning contextual relations, in the form of relative locations between different video regions. We begin by over-segmenting the videos into supervoxels, which have the ability to preserve action boundaries and also reduce the complexity of the problem. Context relations are learned during training which capture displacements from all the supervoxels in a video to those belonging to foreground actions. Then, given a testing video, we select a supervoxel randomly and use the context information acquired during training to estimate the probability of each supervoxel belonging to the foreground action. The walk proceeds to a new supervoxel and the process is repeated for a few steps. This “context walk” generates a conditional distribution of an action over all the supervoxels. A Conditional Random Field is then used to find action proposals in the video, whose confidences are obtained using SVMs. We validated the proposed approach on several datasets and show that context in the form of relative displacements between supervoxels can be extremely useful for action localization. This also results in significantly fewer evaluations of the classifier, in sharp contrast to the alternate sliding window approaches.
Similar papers:
  • Action Detection by Implicit Intentional Motion Clustering [pdf] - Wei Chen, Jason J. Corso
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
  • Action Recognition by Hierarchical Mid-Level Action Elements [pdf] - Tian Lan, Yuke Zhu, Amir Roshan Zamir, Silvio Savarese
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf]
Bilge Soran, Ali Farhadi, Linda Shapiro

Abstract: We all have experienced forgetting habitual actions among our daily activities. For example, we probably have forgotten to turn the lights off before leaving a room or turn the stove off after cooking. In this paper, we propose a solution to the problem of issuing notifications on actions that may be missed. This involves learning about interdependencies between actions and being able to predict an ongoing action while segmenting the input video stream. In order to show a proof of concept, we collected a new egocentric dataset, in which people wear a camera while making lattes 1 . We show promising results on the extremely challeng1A latte is a coffee drink made with espresso and steamed milk [1]. ing task of issuing correct and timely reminders. We also show that our model reliably segments the actions, while predicting the ongoing one when only a few frames from the beginning of the action are observed. The overall prediction accuracy is 46.2% when only 10 frames of an action are seen (2/3 of a sec). Moreover, the overall recognition and segmentation accuracy is shown to be 72.7% when the whole activity sequence is observed. Finally, the online prediction and segmentation accuracy is 68.3% when the prediction is made at every time step.
Similar papers:
  • Action Recognition by Hierarchical Mid-Level Action Elements [pdf] - Tian Lan, Yuke Zhu, Amir Roshan Zamir, Silvio Savarese
  • Action Localization in Videos Through Context Walk [pdf] - Khurram Soomro, Haroon Idrees, Mubarak Shah
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
Entropy Minimization for Convex Relaxation Approaches [pdf]
Mohamed Souiai, Martin R. Oswald, Youngwook Kee, Junmo Kim, Marc Pollefeys, Daniel Cremers

Abstract: Despite their enormous success in solving hard combinatorial problems, convex relaxation approaches often suffer from the fact that the computed solutions are far from binary and that subsequent heuristic binarization may substantially degrade the quality of computed solutions. In this paper, we propose a novel relaxation technique which incorporates the entropy of the objective variable as a measure of relaxation tightness. We show both theoretically and experimentally that augmenting the objective function with an entropy term gives rise to more binary solutions and consequently solutions with a substantially lower optimality gap. We use difference of convex function (DC) programming as an efficient and provably convergent solver for the arising convex-concave minimization problem. We evaluate this approach on three prominent non-convex computer vision challenges: multi-label inpainting, image segmentation and spatio-temporal multi-view reconstruction. These experiments show that our approach consistently yields better solutions with respect to the original integral optimization problem.
Similar papers:
  • Live Repetition Counting [pdf] - Ofir Levy, Lior Wolf
  • Higher-Order Inference for Multi-Class Log-Supermodular Models [pdf] - Jian Zhang, Josip Djolonga, Andreas Krause
  • Exploring Causal Relationships in Visual Object Tracking [pdf] - Karel Lebeda, Simon Hadfield, Richard Bowden
  • Entropy-Based Latent Structured Output Prediction [pdf] - Diane Bouchacourt, Sebastian Nowozin, M. Pawan Kumar
  • Volumetric Bias in Segmentation and Reconstruction: Secrets and Solutions [pdf] - Yuri Boykov, Hossam Isack, Carl Olsson, Ismail Ben Ayed
Oriented Light-Field Windows for Scene Flow [pdf]
Pratul P. Srinivasan, Michael W. Tao, Ren Ng, Ravi Ramamoorthi

Abstract: 2D spatial image windows are used for comparing pixel values in computer vision applications such as correspondence for optical flow and 3D reconstruction, bilateral filtering, and image segmentation. However, pixel window comparisons can suffer from varying defocus blur and perspective at different depths, and can also lead to a loss of precision. In this paper, we leverage the recent use of light-field cameras to propose alternative oriented lightfield windows that enable more robust and accurate pixel comparisons. For Lambertian surfaces focused to the correct depth, the 2D distribution of angular rays from a pixel remains consistent. We build on this idea to develop an oriented 4D light-field window that accounts for shearing (depth), translation (matching), and windowing. Our main application is to scene flow, a generalization of optical flow to the 3D vector field describing the motion of each point in the scene. We show significant benefits of oriented lightfield windows over standard 2D spatial windows. We also demonstrate additional applications of oriented light-field windows for bilateral filtering and image segmentation.
Similar papers:
  • SPM-BP: Sped-up PatchMatch Belief Propagation for Continuous MRFs [pdf] - Yu Li, Dongbo Min, Michael S. Brown, Minh N. Do, Jiangbo Lu
  • Face Flow [pdf] - Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou
  • FlowNet: Learning Optical Flow With Convolutional Networks [pdf] - Alexey Dosovitskiy, Philipp Fischer, Eddy Ilg, Philip Hausser, Caner Hazirbas, Vladimir Golkov, Patrick van der Smagt, Daniel Cremers, Thomas Brox
  • Large Displacement 3D Scene Flow With Occlusion Reasoning [pdf] - Andrei Zanfir, Cristian Sminchisescu
  • Flow Fields: Dense Correspondence Fields for Highly Accurate Large Displacement Optical Flow Estimation [pdf] - Christian Bailer, Bertram Taetz, Didier Stricker
Model-Based Tracking at 300Hz Using Raw Time-of-Flight Observations [pdf]
Jan Stuhmer, Sebastian Nowozin, Andrew Fitzgibbon, Richard Szeliski, Travis Perry, Sunil Acharya, Daniel Cremers, Jamie Shotton

Abstract: modulated, clustered Microsoft modulated, equispaced unmodulated, equispaced our approach high frame-rate frame 0 time Consumer depth cameras have dramatically improved our ability to track rigid, articulated, and deformable 3D objects in real-time. However, depth cameras have a limited temporal resolution (frame-rate) that restricts the accuracy and robustness of tracking, especially for fast or unpredictable motion. In this paper, we show how to perform model-based object tracking which allows to reconstruct the objects depth at an order of magnitude higher frame-rate through simple modifications to an off-the-shelf depth camera. We focus on phase-based time-of-flight (ToF) sensing, which reconstructs each low frame-rate depth image from a set of short exposure raw infrared captures. These raw captures are taken in quick succession near the beginning of each depth frame, and differ in the modulation of their active illumination. We make two contributions. First, we detail how to perform model-based tracking against these raw captures. Second, we show that by reprogramming the camera to space the raw captures uniformly in time, we obtain a 10x higher frame-rate, and thereby improve the ability to track fast-moving objects. inferred depth frame 1 frame 2 phase-based ToF
Similar papers:
  • Joint Probabilistic Data Association Revisited [pdf] - Seyed Hamid Rezatofighi, Anton Milan, Zhen Zhang, Qinfeng Shi, Anthony Dick, Ian Reid
  • Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf] - Naejin Kong, Michael J. Black
  • High Quality Structure From Small Motion for Rolling Shutter Cameras [pdf] - Sunghoon Im, Hyowon Ha, Gyeongmin Choe, Hae-Gon Jeon, Kyungdon Joo, In So Kweon
  • A Versatile Learning-Based 3D Temporal Tracker: Scalable, Robust, Online [pdf] - David Joseph Tan, Federico Tombari, Slobodan Ilic, Nassir Navab
  • Robust Non-Rigid Motion Tracking and Surface Reconstruction Using L0 Regularization [pdf] - Kaiwen Guo, Feng Xu, Yangang Wang, Yebin Liu, Qionghai Dai
3D-Assisted Feature Synthesis for Novel Views of an Object [pdf]
Hao Su, Fan Wang, Eric Yi, Leonidas J. Guibas

Abstract: Comparing two images from different views has been a long-standing challenging problem in computer vision, as visual features are not stable under large view point changes. In this paper, given a single input image of an object, we synthesize its features for other views, leveraging an existing modestly-sized 3D model collection of related but not identical objects.To accomplish this, we study the relationship of image patches between different views of the same object, seeking what we call surrogate patches patches in one view whose feature content predicts well the features of a patch in another view. Based upon these surrogate relationships, we can create feature sets for all views of the latent object on a per patch basis, providing us an augmented multi-view representation of the object. We provide theoretical and empirical analysis of the feature synthesis process, and evaluate the augmented features in fine-grained image retrieval/recognition and instance retrieval tasks. Experimental results show that our synthesized features do enable view-independent comparison between images and perform significantly better than other traditional approaches in this respect.
Similar papers:
  • One Shot Learning via Compositions of Meaningful Patches [pdf] - Alex Wong, Alan L. Yuille
  • Learning Parametric Distributions for Image Super-Resolution: Where Patch Matching Meets Sparse Coding [pdf] - Yongbo Li, Weisheng Dong, Guangming Shi, Xuemei Xie
  • Render for CNN: Viewpoint Estimation in Images Using CNNs Trained With Rendered 3D Model Views [pdf] - Hao Su, Charles R. Qi, Yangyan Li, Leonidas J. Guibas
  • Local Convolutional Features With Unsupervised Training for Image Retrieval [pdf] - Mattis Paulin, Matthijs Douze, Zaid Harchaoui, Julien Mairal, Florent Perronin, Cordelia Schmid
  • Multi-View Convolutional Neural Networks for 3D Shape Recognition [pdf] - Hang Su, Subhransu Maji, Evangelos Kalogerakis, Erik Learned-Miller
Contour Guided Hierarchical Model for Shape Matching [pdf]
Yuanqi Su, Yuehu Liu, Bonan Cuan, Nanning Zheng

Abstract: For its simplicity and effectiveness, star model is popular in shape matching. However, it suffers from the loose geometric connections among parts. In the paper, we present a novel algorithm that reconsiders these connections and reduces the global matching to a set of interrelated local matching. For the purpose, we divide the shape template into overlapped parts and model the matching through a part-based layered structure that uses the latent variable to constrain parts deformation. As for inference, each part is used for localizing candidates by the partial matching. Thanks to the contour fragments, the partial matching can be solved via modified dynamic programming. The overlapped regions among parts of the template are then explored to make the candidates of parts meet at their shared points. The process is fulfilled via a refined procedure based on iterative dynamic programming. Results on ETHZ shape and Inria Horse datasets demonstrate the benefits of the proposed algorithm.
Similar papers:
  • Interactive Visual Hull Refinement for Specular and Transparent Object Surface Reconstruction [pdf] - Xinxin Zuo, Chao Du, Sen Wang, Jiangbin Zheng, Ruigang Yang
  • BodyPrint: Pose Invariant 3D Shape Matching of Human Bodies [pdf] - Jiangping Wang, Kai Ma, Vivek Kumar Singh, Thomas Huang, Terrence Chen
  • Compositional Hierarchical Representation of Shape Manifolds for Classification of Non-Manifold Shapes [pdf] - Mete Ozay, Umit Rusen Aktas, Jeremy L. Wyatt, Ales Leonardis
  • Contour Flow: Middle-Level Motion Estimation by Combining Motion Segmentation and Contour Alignment [pdf] - Huijun Di, Qingxuan Shi, Feng Lv, Ming Qin, Yao Lu
  • 3D Fragment Reassembly Using Integrated Template Guidance and Fracture-Region Matching [pdf] - Kang Zhang, Wuyi Yu, Mary Manhein, Warren Waggenspack, Xin Li
Multi-Task Learning With Low Rank Attribute Embedding for Person Re-Identification [pdf]
Chi Su, Fan Yang, Shiliang Zhang, Qi Tian, Larry S. Davis, Wen Gao

Abstract: We propose a novel Multi-Task Learning with Low Rank Attribute Embedding (MTL-LORAE) framework for person re-identification. Re-identifications from multiple cameras are regarded as related tasks to exploit shared information to improve re-identification accuracy. Both low level features and semantic/data-driven attributes are utilized. Since attributes are generally correlated, we introduce a low rank attribute embedding into the MTL formulation to embed original binary attributes to a continuous attribute space, where incorrect and incomplete attributes are rectified and recovered to better describe people. The learning objective function consists of a quadratic loss regarding class labels and an attribute embedding error, which is solved by an alternating optimization procedure. Experiments on four person re-identification datasets have demonstrated that MTLLORAE outperforms existing approaches by a large margin and produces promising results.
Similar papers:
  • Attributed Grammars for Joint Estimation of Human Attributes, Part and Pose [pdf] - Seyoung Park, Song-Chun Zhu
  • Discovering the Spatial Extent of Relative Attributes [pdf] - Fanyi Xiao, Yong Jae Lee
  • Just Noticeable Differences in Visual Attributes [pdf] - Aron Yu, Kristen Grauman
  • Learning Deep Representation With Large-Scale Attributes [pdf] - Wanli Ouyang, Hongyang Li, Xingyu Zeng, Xiaogang Wang
  • A Unified Multiplicative Framework for Attribute Learning [pdf] - Kongming Liang, Hong Chang, Shiguang Shan, Xilin Chen
Multi-View Convolutional Neural Networks for 3D Shape Recognition [pdf]
Hang Su, Subhransu Maji, Evangelos Kalogerakis, Erik Learned-Miller

Abstract: A longstanding question in computer vision concerns the representation of 3D shapes for recognition: should 3D shapes be represented with descriptors operating on their native 3D formats, such as voxel grid or polygon mesh, or can they be effectively represented with view-based descriptors? We address this question in the context of learning to recognize 3D shapes from a collection of their rendered views on 2D images. We first present a standard CNN architecture trained to recognize the shapes rendered views independently of each other, and show that a 3D shape can be recognized even from a single view at an accuracy far higher than using state-of-the-art 3D shape descriptors. Recognition rates further increase when multiple views of the shapes are provided. In addition, we present a novel CNN architecture that combines information from multiple views of a 3D shape into a single and compact shape descriptor offering even better recognition performance. The same architecture can be applied to accurately recognize human hand-drawn sketches of shapes. We conclude that a collection of 2D views can be highly informative for 3D shape recognition and is amenable to emerging CNN architectures and their derivatives.
Similar papers:
  • Local Convolutional Features With Unsupervised Training for Image Retrieval [pdf] - Mattis Paulin, Matthijs Douze, Zaid Harchaoui, Julien Mairal, Florent Perronin, Cordelia Schmid
  • Aggregating Local Deep Features for Image Retrieval [pdf] - Artem Babenko, Victor Lempitsky
  • BodyPrint: Pose Invariant 3D Shape Matching of Human Bodies [pdf] - Jiangping Wang, Kai Ma, Vivek Kumar Singh, Thomas Huang, Terrence Chen
  • 3D-Assisted Feature Synthesis for Novel Views of an Object [pdf] - Hao Su, Fan Wang, Eric Yi, Leonidas J. Guibas
  • Compositional Hierarchical Representation of Shape Manifolds for Classification of Non-Manifold Shapes [pdf] - Mete Ozay, Umit Rusen Aktas, Jeremy L. Wyatt, Ales Leonardis
Render for CNN: Viewpoint Estimation in Images Using CNNs Trained With Rendered 3D Model Views [pdf]
Hao Su, Charles R. Qi, Yangyan Li, Leonidas J. Guibas

Abstract: Object viewpoint estimation from 2D images is an essential task in computer vision. However, two issues hinder its progress: scarcity of training data with viewpoint annotations, and a lack of powerful features. Inspired by the growing availability of 3D models, we propose a framework to address both issues by combining render-based image synthesis and CNNs (Convolutional Neural Networks). We believe that 3D models have the potential in generating a large number of images of high variation, which can be well exploited by deep CNN with a high learning capacity. Towards this goal, we propose a scalable and overfitresistant image synthesis pipeline, together with a novel CNN specifically tailored for the viewpoint estimation task. Experimentally, we show that the viewpoint estimation from our pipeline can significantly outperform state-of-the-art methods on PASCAL 3D+ benchmark.
Similar papers:
  • Real-Time Pose Estimation Piggybacked on Object Detection [pdf] - Roman Juranek, Adam Herout, Marketa Dubska, Pavel Zemcik
  • Continuous Pose Estimation With a Spatial Ensemble of Fisher Regressors [pdf] - Michele Fenzi, Laura Leal-Taixe, Jorn Ostermann, Tinne Tuytelaars
  • Single Image Pop-Up From Discriminatively Learned Parts [pdf] - Menglong Zhu, Xiaowei Zhou, Kostas Daniilidis
  • Scene-Domain Active Part Models for Object Representation [pdf] - Zhou Ren, Chaohui Wang, Alan L. Yuille
  • Unsupervised Generation of a Viewpoint Annotated Car Dataset From Videos [pdf] - Nima Sedaghat, Thomas Brox
Discriminative Low-Rank Tracking [pdf]
Yao Sui, Yafei Tang, Li Zhang

Abstract: Good tracking performance is in general attributed to accurate representation over previously obtained targets or reliable discrimination between the target and the surrounding background. In this work, we exploit the advantages of the both approaches to achieve a robust tracker. We construct a subspace to represent the target and the neighboring background, and simultaneously propagate their class labels via the learned subspace. Moreover, we propose a novel criterion to identify the target from numerous target candidates on each frame, which takes into account both discrimination reliability and representation accuracy. In addition, with the proposed criterion, the ambiguity in the class labels of the neighboring background samples, which often influences the reliability of discriminative tracking model, is effectively alleviated, while the training set is still kept small. Extensive experiments demonstrate that our tracker performs favourably against many other state-of-the-art trackers.
Similar papers:
  • Local Subspace Collaborative Tracking [pdf] - Lin Ma, Xiaoqin Zhang, Weiming Hu, Junliang Xing, Jiwen Lu, Jie Zhou
  • SOWP: Spatially Ordered and Weighted Patch Descriptor for Visual Tracking [pdf] - Han-Ul Kim, Dae-Youn Lee, Jae-Young Sim, Chang-Su Kim
  • Linearization to Nonlinear Learning for Visual Tracking [pdf] - Bo Ma, Hongwei Hu, Jianbing Shen, Yuping Zhang, Fatih Porikli
  • Learning to Track: Online Multi-Object Tracking by Decision Making [pdf] - Yu Xiang, Alexandre Alahi, Silvio Savarese
  • Visual Tracking With Fully Convolutional Networks [pdf] - Lijun Wang, Wanli Ouyang, Xiaogang Wang, Huchuan Lu
Listening With Your Eyes: Towards a Practical Visual Speech Recognition System Using Deep Boltzmann Machines [pdf]
Chao Sui, Mohammed Bennamoun, Roberto Togneri

Abstract: This paper presents a novel feature learning method for visual speech recognition using Deep Boltzmann Machines (DBM). Unlike all existing visual feature extraction techniques which solely extract features from video sequences, our method is able to explore both acoustic information and visual information to learn a better visual feature representation. During the test stage, instead of using both audio and visual signals, only the videos are used to generate the missing audio features, and both the given visual and audio features are used to produce a joint representation. We carried out experiments on a new large scale audio-visual corpus, and experimental results show that our proposed techniques outperform the performance of handcrafted features and previously learned features and can be adopted by other deep learning systems.
Similar papers:
  • Selecting Relevant Web Trained Concepts for Automated Event Retrieval [pdf] - Bharat Singh, Xintong Han, Zhe Wu, Vlad I. Morariu, Larry S. Davis
  • Weakly-Supervised Alignment of Video With Text [pdf] - Piotr Bojanowski, Remi Lajugie, Edouard Grave, Francis Bach, Ivan Laptev, Jean Ponce, Cordelia Schmid
  • Unsupervised Extraction of Video Highlights Via Robust Recurrent Auto-Encoders [pdf] - Huan Yang, Baoyuan Wang, Stephen Lin, David Wipf, Minyi Guo, Baining Guo
  • Contractive Rectifier Networks for Nonlinear Maximum Margin Classification [pdf] - Senjian An, Munawar Hayat, Salman H. Khan, Mohammed Bennamoun, Farid Boussaid, Ferdous Sohel
  • FaceDirector: Continuous Control of Facial Performance in Video [pdf] - Charles Malleson, Jean-Charles Bazin, Oliver Wang, Derek Bradley, Thabo Beeler, Adrian Hilton, Alexander Sorkine-Hornung
Automatic Concept Discovery From Parallel Text and Visual Corpora [pdf]
Chen Sun, Chuang Gan, Ram Nevatia

Abstract: Humans connect language and vision to perceive the world. How to build a similar connection for computers? One possible way is via visual concepts, which are text terms that relate to visually discriminative entities. We propose an automatic visual concept discovery algorithm using parallel text and visual corpora; it filters text terms based on the visual discriminative power of the associated images, and groups them into concepts using visual and semantic similarities. We illustrate the applications of the discovered concepts using bidirectional image and sentence retrieval task and image tagging task, and show that the discovered concepts not only outperform several large sets of manually selected concepts significantly, but also achieves the stateof-the-art performance in the retrieval task.
Similar papers:
  • Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books [pdf] - Yukun Zhu, Ryan Kiros, Rich Zemel, Ruslan Salakhutdinov, Raquel Urtasun, Antonio Torralba, Sanja Fidler
  • Multimodal Convolutional Neural Networks for Matching Image and Sentence [pdf] - Lin Ma, Zhengdong Lu, Lifeng Shang, Hang Li
  • LEWIS: Latent Embeddings for Word Images and their Semantics [pdf] - Albert Gordo, Jon Almazan, Naila Murray, Florent Perronin
  • Learning Like a Child: Fast Novel Visual Concept Learning From Sentence Descriptions of Images [pdf] - Junhua Mao, Xu Wei, Yi Yang, Jiang Wang, Zhiheng Huang, Alan L. Yuille
  • Selecting Relevant Web Trained Concepts for Automated Event Retrieval [pdf] - Bharat Singh, Xintong Han, Zhe Wu, Vlad I. Morariu, Larry S. Davis
Human Action Recognition Using Factorized Spatio-Temporal Convolutional Networks [pdf]
Lin Sun, Kui Jia, Dit-Yan Yeung, Bertram E. Shi

Abstract: Human actions in video sequences are threedimensional (3D) spatio-temporal signals characterizing both the visual appearance and motion dynamics of the involved humans and objects. Inspired by the success of convolutional neural networks (CNN) for image classification, recent attempts have been made to learn 3D CNNs for recognizing human actions in videos. However, partly due to the high complexity of training 3D convolution kernels and the need for large quantities of training videos, only limited success has been reported. This has triggered us to investigate in this paper a new deep architecture which can handle 3D signals more effectively. Specifically, we propose factorized spatio-temporal convolutional networks (FST CN) that factorize the original 3D convolution kernel learning as a sequential process of learning 2D spatial kernels in the lower layers (called spatial convolutional layers), followed by learning 1D temporal kernels in the upper layers (called temporal convolutional layers). We introduce a novel transformation and permutation operator to make factorization in FST CN possible. Moreover, to address the issue of sequence alignment, we propose an effective training and inference strategy based on sampling multiple video clips from a given action video sequence. We have tested FST CN on two commonly used benchmark datasets (UCF-101 and HMDB-51). Without using auxiliary training videos to boost the performance, FST CN outperforms existing CNN based methods and a
Similar papers:
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Action Recognition by Hierarchical Mid-Level Action Elements [pdf] - Tian Lan, Yuke Zhu, Amir Roshan Zamir, Silvio Savarese
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Learning Spatiotemporal Features With 3D Convolutional Networks [pdf] - Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, Manohar Paluri
Depth-Based Hand Pose Estimation: Data, Methods, and Challenges [pdf]
James S. Supancic III, Gregory Rogez, Yi Yang, Jamie Shotton, Deva Ramanan

Abstract: Hand pose estimation has matured rapidly in recent years. The introduction of commodity depth sensors and a multitude of practical applications have spurred new advances. We provide an extensive analysis of the state-of-theart, focusing on hand pose estimation from a single depth frame. To do so, we have implemented a considerable number of systems, and will release all software and evaluation code. We summarize important conclusions here: (1) Pose estimation appears roughly solved for scenes with isolated hands. However, methods still struggle to analyze cluttered scenes where hands may be interacting with nearby objects and surfaces. To spur further progress we introduce a challenging new dataset with diverse, cluttered scenes. (2) Many methods evaluate themselves with disparate criteria, making comparisons difficult. We define a consistent evaluation criteria, rigorously motivated by human experiments. (3) We introduce a simple nearest-neighbor baseline that outperforms most existing systems. This implies that most systems do not generalize beyond their training sets. This also reinforces the under-appreciated point that training data is as important as the model itself. We conclude with directions for future progress.
Similar papers:
  • Lending A Hand: Detecting Hands and Recognizing Activities in Complex Egocentric Interactions [pdf] - Sven Bambach, Stefan Lee, David J. Crandall, Chen Yu
  • 3D Hand Pose Estimation Using Randomized Decision Forest With Segmentation Index Points [pdf] - Peiyi Li, Haibin Ling, Xi Li, Chunyuan Liao
  • Training a Feedback Loop for Hand Pose Estimation [pdf] - Markus Oberweger, Paul Wohlhart, Vincent Lepetit
  • Understanding Everyday Hands in Action From RGB-D Images [pdf] - Gregory Rogez, James S. Supancic III, Deva Ramanan
  • A Collaborative Filtering Approach to Real-Time Hand Pose Estimation [pdf] - Chiho Choi, Ayan Sinha, Joon Hee Choi, Sujin Jang, Karthik Ramani
What Makes Tom Hanks Look Like Tom Hanks [pdf]
Supasorn Suwajanakorn, Steven M. Seitz, Ira Kemelmacher-Shlizerman

Abstract: We reconstruct a controllable model of a person from a large photo collection that captures his or her persona, i.e., physical appearance and behavior. The ability to operate on unstructured photo collections enables modeling a huge number of people, including celebrities and other well photographed people without requiring them to be scanned. Moreover, we show the ability to drive or puppeteer the captured person B using any other video of a different person A. In this scenario, B acts out the role of person A, but retains his/her own personality and character. Our system is based on a novel combination of 3D face reconstruction, tracking, alignment, and multi-texture modeling, applied to the puppeteering problem. We demonstrate convincing results on a large variety of celebrities derived from Internet imagery and video.
Similar papers:
  • Person Recognition in Personal Photo Collections [pdf] - Seong Joon Oh, Rodrigo Benenson, Mario Fritz, Bernt Schiele
  • From Emotions to Action Units With Hidden and Semi-Hidden-Task Learning [pdf] - Adria Ruiz, Joost Van de Weijer, Xavier Binefa
  • FaceDirector: Continuous Control of Facial Performance in Video [pdf] - Charles Malleson, Jean-Charles Bazin, Oliver Wang, Derek Bradley, Thabo Beeler, Adrian Hilton, Alexander Sorkine-Hornung
  • Face Flow [pdf] - Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou
  • Example-Based Modeling of Facial Texture From Deficient Data [pdf] - Arnaud Dessein, William A. P. Smith, Richard C. Wilson, Edwin R. Hancock
Optimizing the Viewing Graph for Structure-From-Motion [pdf]
Chris Sweeney, Torsten Sattler, Tobias Hollerer, Matthew Turk, Marc Pollefeys

Abstract: The viewing graph represents a set of views that are related by pairwise relative geometries. In the context of Structure-from-Motion (SfM), the viewing graph is the input to the incremental or global estimation pipeline. Much effort has been put towards developing robust algorithms to overcome potentially inaccurate relative geometries in the viewing graph during SfM. In this paper, we take a fundamentally different approach to SfM and instead focus on improving the quality of the viewing graph before applying SfM. Our main contribution is a novel optimization that improves the quality of the relative geometries in the viewing graph by enforcing loop consistency constraints with the epipolar point transfer. We show that this optimization greatly improves the accuracy of relative poses in the viewing graph and removes the need for filtering steps or robust algorithms typically used in global SfM methods. In addition, the optimized viewing graph can be used to efficiently calibrate cameras at scale. We combine our viewing graph optimization and focal length calibration into a global SfM pipeline that is more efficient than existing approaches. To our knowledge, ours is the first global SfM pipeline capable of handling uncalibrated image sets.
Similar papers:
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Non-Parametric Structure-Based Calibration of Radially Symmetric Cameras [pdf] - Federico Camposeco, Torsten Sattler, Marc Pollefeys
  • Extended Depth of Field Catadioptric Imaging Using Focal Sweep [pdf] - Ryunosuke Yokoya, Shree K. Nayar
  • Structure From Motion Using Structure-Less Resection [pdf] - Enliang Zheng, Changchang Wu
  • Global Structure-From-Motion by Similarity Averaging [pdf] - Zhaopeng Cui, Ping Tan
Depth Selective Camera: A Direct, On-Chip, Programmable Technique for Depth Selectivity in Photography [pdf]
Ryuichi Tadano, Adithya Kumar Pediredla, Ashok Veeraraghavan

Abstract: Time of flight (ToF) cameras use a temporally modulated light source and measure correlation between the reflected light and a sensor modulation pattern, in order to infer scene depth. In this paper, we show that such correlational sensors can also be used to selectively accept or reject light rays from certain scene depths. The basic idea is to carefully select illumination and sensor modulation patterns such that the correlation is non-zero only in the selected depth range thus light reflected from objects outside this depth range do not affect the correlational measurements. We demonstrate a prototype depth-selective camera and highlight two potential applications: imaging through scattering media and virtual blue screening. This depthselectivity can be used to reject back-scattering and reflection from media in front of the subjects of interest, thereby significantly enhancing the ability to image through scattering media- critical for applications such as car navigation in fog and rain. Similarly, such depth selectivity can also be utilized as a virtual blue-screen in cinematography by rejecting light reflecting from background, while selectively retaining light contributions from the foreground subject.
Similar papers:
  • Photometric Stereo With Small Angular Variations [pdf] - Jian Wang, Yasuyuki Matsushita, Boxin Shi, Aswin C. Sankaranarayanan
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Nighttime Haze Removal With Glow and Multiple Light Colors [pdf] - Yu Li, Robby T. Tan, Michael S. Brown
  • Photometric Stereo in a Scattering Medium [pdf] - Zak Murez, Tali Treibitz, Ravi Ramamoorthi, David Kriegman
  • SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf] - Kensei Jo, Mohit Gupta, Shree K. Nayar
A Versatile Learning-Based 3D Temporal Tracker: Scalable, Robust, Online [pdf]
David Joseph Tan, Federico Tombari, Slobodan Ilic, Nassir Navab

Abstract: This paper proposes a temporal tracking algorithm based on Random Forest that uses depth images to estimate and track the 3D pose of a rigid object in real-time. Compared to the state of the art aimed at the same goal, our algorithm holds important attributes such as high robustness against holes and occlusion, low computational cost of both learning and tracking stages, and low memory consumption. These are obtained (a) by a novel formulation of the learning strategy, based on a dense sampling of the camera viewpoints and learning independent trees from a single image for each camera view; as well as, (b) by an insightful occlusion handling strategy that enforces the forest to recognize the objects local and global structures. Due to these attributes, we report state-of-the-art tracking accuracy on benchmark datasets, and accomplish remarkable scalability with the number of targets, being able to simultaneously track the pose of over a hundred objects at 30 fps with an off-the-shelf CPU. In addition, the fast learning time enables us to extend our algorithm as a robust online tracker for model-free 3D objects under different viewpoints and appearance changes as demonstrated by the experiments.
Similar papers:
  • Learning to Track: Online Multi-Object Tracking by Decision Making [pdf] - Yu Xiang, Alexandre Alahi, Silvio Savarese
  • Joint Probabilistic Data Association Revisited [pdf] - Seyed Hamid Rezatofighi, Anton Milan, Zhen Zhang, Qinfeng Shi, Anthony Dick, Ian Reid
  • Understanding and Diagnosing Visual Tracking Systems [pdf] - Naiyan Wang, Jianping Shi, Dit-Yan Yeung, Jiaya Jia
  • Online Object Tracking With Proposal Selection [pdf] - Yang Hua, Karteek Alahari, Cordelia Schmid
  • Model-Based Tracking at 300Hz Using Raw Time-of-Flight Observations [pdf] - Jan Stuhmer, Sebastian Nowozin, Andrew Fitzgibbon, Richard Szeliski, Travis Perry, Sunil Acharya, Daniel Cremers, Jamie Shotton
Improving Image Classification With Location Context [pdf]
Kevin Tang, Manohar Paluri, Li Fei-Fei, Rob Fergus, Lubomir Bourdev

Abstract: With the widespread availability of cellphones and cameras that have GPS capabilities, it is common for images being uploaded to the Internet today to have GPS coordinates associated with them. In addition to research that tries to predict GPS coordinates from visual features, this also opens up the door to problems that are conditioned on the availability of GPS coordinates. In this work, we tackle the problem of performing image classification with location context, in which we are given the GPS coordinates for images in both the train and test phases. We explore different ways of encoding and extracting features from the GPS coordinates, and show how to naturally incorporate these features into a Convolutional Neural Network (CNN), the current state-of-the-art for most image classification and recognition problems. We also show how it is possible to simultaneously learn the optimal pooling radii for a subset of our features within the CNN framework. To evaluate our model and to help promote research in this area, we identify a set of location-sensitive concepts and annotate a subset of the Yahoo Flickr Creative Commons 100M dataset that has GPS coordinates with these concepts, which we make publicly available. By leveraging location context, we are able to achieve almost a 7% gain in mean average precision.
Similar papers:
  • RGB-W: When Vision Meets Wireless [pdf] - Alexandre Alahi, Albert Haque, Li Fei-Fei
  • Human Parsing With Contextualized Convolutional Neural Network [pdf] - Xiaodan Liang, Chunyan Xu, Xiaohui Shen, Jianchao Yang, Si Liu, Jinhui Tang, Liang Lin, Shuicheng Yan
  • Webly Supervised Learning of Convolutional Networks [pdf] - Xinlei Chen, Abhinav Gupta
  • Multi-Scale Recognition With DAG-CNNs [pdf] - Songfan Yang, Deva Ramanan
  • Context Aware Active Learning of Activity Recognition Models [pdf] - Mahmudul Hasan, Amit K. Roy-Chowdhury
Multi-Kernel Correlation Filter for Visual Tracking [pdf]
Ming Tang, Jiayi Feng

Abstract: Correlation filter based trackers are ranked top in terms of performances. Nevertheless, they only employ a single kernel at a time. In this paper, we will derive a multikernel correlation filter (MKCF) based tracker which fully takes advantage of the invariance-discriminative power spectrums of various features to further improve the performance. Moreover, it may easily introduce location and representation errors to search several discrete scales for the proper one of the object bounding box, because normally the discrete candidate scales are determined and the corresponding feature pyramid are generated ahead of searching. In this paper, we will propose a novel and efficient scale estimation method based on optimal bisection search and fast evaluation of features. Our scale estimation method is the first one that uses the truly minimal number of layers of feature pyramid and avoids constructing the pyramid before searching for proper scales.
Similar papers:
  • Online Object Tracking With Proposal Selection [pdf] - Yang Hua, Karteek Alahari, Cordelia Schmid
  • SOWP: Spatially Ordered and Weighted Patch Descriptor for Visual Tracking [pdf] - Han-Ul Kim, Dae-Youn Lee, Jae-Young Sim, Chang-Su Kim
  • Understanding and Diagnosing Visual Tracking Systems [pdf] - Naiyan Wang, Jianping Shi, Dit-Yan Yeung, Jiaya Jia
  • Hierarchical Convolutional Features for Visual Tracking [pdf] - Chao Ma, Jia-Bin Huang, Xiaokang Yang, Ming-Hsuan Yang
  • Learning Spatially Regularized Correlation Filters for Visual Tracking [pdf] - Martin Danelljan, Gustav Hager, Fahad Shahbaz Khan, Michael Felsberg
Opening the Black Box: Hierarchical Sampling Optimization for Estimating Human Hand Pose [pdf]
Danhang Tang, Jonathan Taylor, Pushmeet Kohli, Cem Keskin, Tae-Kyun Kim, Jamie Shotton

Abstract: We address the problem of hand pose estimation, formulated as an inverse problem. Typical approaches optimize an energy function over pose parameters using a black box image generation procedure. This procedure knows little about either the relationships between the parameters or the form of the energy function. In this paper, we show that we can significantly improve upon black box optimization by exploiting high-level knowledge of the parameter structure and using a local surrogate energy function. Our new framework, hierarchical sampling optimization, consists of a sequence of predictors organized into a kinematic hierarchy. Each predictor is conditioned on its ancestors, and generates a set of samples over a subset of the pose parameters. The highly-efficient surrogate energy is used to select among samples. Having evaluated the full hierarchy, the partial pose samples are concatenated to generate a full-pose hypothesis. Several hypotheses are generated using the same procedure, and finally the original full energy function selects the best result. Experimental evaluation on three publically available datasets show that our method is particularly impressive in low-compute scenarios where it significantly outperforms all other state-of-the-art methods.
Similar papers:
  • Understanding Everyday Hands in Action From RGB-D Images [pdf] - Gregory Rogez, James S. Supancic III, Deva Ramanan
  • 3D Hand Pose Estimation Using Randomized Decision Forest With Segmentation Index Points [pdf] - Peiyi Li, Haibin Ling, Xi Li, Chunyuan Liao
  • Depth-Based Hand Pose Estimation: Data, Methods, and Challenges [pdf] - James S. Supancic III, Gregory Rogez, Yi Yang, Jamie Shotton, Deva Ramanan
  • Training a Feedback Loop for Hand Pose Estimation [pdf] - Markus Oberweger, Paul Wohlhart, Vincent Lepetit
  • A Collaborative Filtering Approach to Real-Time Hand Pose Estimation [pdf] - Chiho Choi, Ayan Sinha, Joon Hee Choi, Sujin Jang, Karthik Ramani
Secrets of GrabCut and Kernel K-Means [pdf]
Meng Tang, Ismail Ben Ayed, Dmitrii Marin, Yuri Boykov

Abstract: The log-likelihood energy term in popular model-fitting segmentation methods, e.g. [39, 8, 28, 10], is presented as a generalized probabilistic K-means energy [16] for color space clustering. This interpretation reveals some limitations, e.g. over-fitting. We propose an alternative approach to color clustering using kernel K-means energy with wellknown properties such as non-linear separation and scalability to higher-dimensional feature spaces. Our bound formulation for kernel K-means allows to combine general pair-wise feature clustering methods with image grid regularization using graph cuts, similarly to standard color model fitting techniques for segmentation. Unlike histogram or GMM fitting [39, 28], our approach is closely related to average association and normalized cut. But, in contrast to previous pairwise clustering algorithms, our approach can incorporate any standard geometric regularization in the image domain. We analyze extreme cases for kernel bandwidth (e.g. Gini bias) and demonstrate effectiveness of KNN-based adaptive bandwidth strategies. Our kernel K-means approach to segmentation benefits from higherdimensional features where standard model-fitting fails.
Similar papers:
  • Piecewise Flat Embedding for Image Segmentation [pdf] - Yizhou Yu, Chaowei Fang, Zicheng Liao
  • Per-Sample Kernel Adaptation for Visual Recognition and Grouping [pdf] - Borislav Antic, Bjorn Ommer
  • Low Dimensional Explicit Feature Maps [pdf] - Ondrej Chum
  • Volumetric Bias in Segmentation and Reconstruction: Secrets and Solutions [pdf] - Yuri Boykov, Hossam Isack, Carl Olsson, Ismail Ben Ayed
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
Realtime Edge-Based Visual Odometry for a Monocular Camera [pdf]
Juan Jose Tarrio, Sol Pedre

Abstract: In this work we present a novel algorithm for realtime visual odometry for a monocular camera. The main idea is to develop an approach between classical feature-based visual odometry systems and modern direct dense/semi-dense methods, trying to benefit from the best attributes of both. Similar to feature-based systems, we extract information from the images, instead of working with raw image intensities as direct methods. In particular, the information extracted are the edges present in the image, while the rest of the algorithm is designed to take advantage of the structural information provided when pixels are treated as edges. Edge extraction is an efficient and higly parallelizable operation. The edge depth information extracted is dense enough to allow acceptable surface fitting, similar to modern semi-dense methods. This is a valuable attribute that feature-based odometry lacks. Experimental results show that the proposed method has similar drift than state of the art feature-based and direct methods, and is a simple algorithm that runs at realtime and can be parallelized. Finally, we have also developed an inertial aided version that successfully stabilizes an unmanned air vehicle in complex indoor environments using only a frontal camera, while running the complete solution in the embedded hardware on board the vehicle.
Similar papers:
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • Model-Based Tracking at 300Hz Using Raw Time-of-Flight Observations [pdf] - Jan Stuhmer, Sebastian Nowozin, Andrew Fitzgibbon, Richard Szeliski, Travis Perry, Sunil Acharya, Daniel Cremers, Jamie Shotton
  • Occlusion-Aware Depth Estimation Using Light-Field Cameras [pdf] - Ting-Chun Wang, Alexei A. Efros, Ravi Ramamoorthi
  • Holistically-Nested Edge Detection [pdf] - Saining Xie, Zhuowen Tu
  • Robust RGB-D Odometry Using Point and Line Features [pdf] - Yan Lu, Dezhen Song
Cluster-Based Point Set Saliency [pdf]
Flora Ponjou Tasse, Jiri Kosinka, Neil Dodgson

Abstract: We propose a cluster-based approach to point set saliency detection, a challenge since point sets lack topological information. A point set is first decomposed into small clusters, using fuzzy clustering. We evaluate cluster uniqueness and spatial distribution of each cluster and combine these values into a cluster saliency function. Finally, the probabilities of points belonging to each cluster are used to assign a saliency to each point. Our approach detects fine-scale salient features and uninteresting regions consistently have lower saliency values. We evaluate the proposed saliency model by testing our saliency-based keypoint detection against a 3D interest point detection benchmark. The evaluation shows that our method achieves a good balance between false positive and false negative error rates, without using any topological information.
Similar papers:
  • Generic Promotion of Diffusion-Based Salient Object Detection [pdf] - Peng Jiang, Nuno Vasconcelos, Jingliang Peng
  • Learning to Predict Saliency on Face Images [pdf] - Mai Xu, Yun Ren, Zulin Wang
  • HARF: Hierarchy-Associated Rich Features for Salient Object Detection [pdf] - Wenbin Zou, Nikos Komodakis
  • A Data-Driven Metric for Comprehensive Evaluation of Saliency Models [pdf] - Jia Li, Changqun Xia, Yafei Song, Shu Fang, Xiaowu Chen
  • SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf] - Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao
Detection and Segmentation of 2D Curved Reflection Symmetric Structures [pdf]
Ching L. Teo, Cornelia Fermuller, Yiannis Aloimonos

Abstract: Symmetry, as one of the key components of Gestalt theory, provides an important mid-level cue that serves as input to higher visual processes such as segmentation. In this work, we propose a complete approach that links the detection of curved reflection symmetries to produce symmetryconstrained segments of structures/regions in real images with clutter. For curved reflection symmetry detection, we leverage on patch-based symmetric features to train a Structured Random Forest classifier that detects multiscaled curved symmetries in 2D images. Next, using these curved symmetries, we modulate a novel symmetryconstrained foreground-background segmentation by their symmetry scores so that we enforce global symmetrical consistency in the final segmentation. This is achieved by imposing a pairwise symmetry prior that encourages symmetric pixels to have the same labels over a MRF-based representation of the input image edges, and the final segmentation is obtained via graph-cuts. Experimental results over four publicly available datasets containing annotated symmetric structures: 1) SYMMAX-300 [38], 2) BSD-Parts, 3) Weizmann Horse (both from [18]) and 4) NY-roads [35] demonstrate the approachs applicability to different environments with state-of-the-art performance.
Similar papers:
  • Learning to Combine Mid-Level Cues for Object Proposal Generation [pdf] - Tom Lee, Sanja Fidler, Sven Dickinson
  • Structured Feature Selection [pdf] - Tian Gao, Ziheng Wang, Qiang Ji
  • Extraction of Virtual Baselines From Distorted Document Images Using Curvilinear Projection [pdf] - Gaofeng Meng, Zuming Huang, Yonghong Song, Shiming Xiang, Chunhong Pan
  • One Shot Learning via Compositions of Meaningful Patches [pdf] - Alex Wong, Alan L. Yuille
  • Merging the Unmatchable: Stitching Visually Disconnected SfM Models [pdf] - Andrea Cohen, Torsten Sattler, Marc Pollefeys
Deep Learning Strong Parts for Pedestrian Detection [pdf]
Yonglong Tian, Ping Luo, Xiaogang Wang, Xiaoou Tang

Abstract: Recent advances in pedestrian detection are attained by transferring the learned features of Convolutional Neural Network (ConvNet) to pedestrians. This ConvNet is typically pre-trained with massive general object categories (e.g. ImageNet). Although these features are able to handle variations such as poses, viewpoints, and lightings, they may fail when pedestrian images with complex occlusions are present. Occlusion handling is one of the most important problem in pedestrian detection. Unlike previous deep models that directly learned a single detector for pedestrian detection, we propose DeepParts, which consists of extensive part detectors. DeepParts has several appealing properties. First, DeepParts can be trained on weakly labeled data, i.e. only pedestrian bounding boxes without part annotations are provided. Second, DeepParts is able to handle low IoU positive proposals that shift away from ground truth. Third, each part detector in DeepParts is a strong detector that can detect pedestrian by observing only a part of a proposal. Extensive experiments in Caltech dataset demonstrate the effectiveness of DeepParts, which yields a new state-of-the-art miss rate of 11.89%, outperforming the second best method by 10%.
Similar papers:
  • Convolutional Channel Features [pdf] - Bin Yang, Junjie Yan, Zhen Lei, Stan Z. Li
  • Neural Activation Constellations: Unsupervised Part Model Discovery With Convolutional Networks [pdf] - Marcel Simon, Erik Rodner
  • Pedestrian Travel Time Estimation in Crowded Scenes [pdf] - Shuai Yi, Hongsheng Li, Xiaogang Wang
  • Human Pose Estimation in Videos [pdf] - Dong Zhang, Mubarak Shah
  • Beyond Tree Structure Models: A New Occlusion Aware Graphical Model for Human Pose Estimation [pdf] - Lianrui Fu, Junge Zhang, Kaiqi Huang
Text Flow: A Unified Text Detection System in Natural Scene Images [pdf]
Shangxuan Tian, Yifeng Pan, Chang Huang, Shijian Lu, Kai Yu, Chew Lim Tan

Abstract: The prevalent scene text detection approach follows four sequential steps comprising character candidate detection, false character candidate removal, text line extraction, and text line verification. However, errors occur and accumulate throughout each of these sequential steps which often lead to low detection performance. To address these issues, we propose a unified scene text detection system, namely Text Flow, by utilizing the minimum cost (min-cost) flow network model. With character candidates detected by cascade boosting, the min-cost flow network model integrates the last three sequential steps into a single process which solves the error accumulation problem at both character level and text line level effectively. The proposed technique has been tested on three public datasets, i.e, ICDAR2011 dataset, ICDAR2013 dataset and a multilingual dataset and it outperforms the state-of-the-art methods on all three datasets with much higher recall and F-score. The good performance on the multilingual dataset shows that the proposed technique can be used for the detection of texts in different languages.
Similar papers:
  • Learning Common Sense Through Visual Abstraction [pdf] - Ramakrishna Vedantam, Xiao Lin, Tanmay Batra, C. Lawrence Zitnick, Devi Parikh
  • Extraction of Virtual Baselines From Distorted Document Images Using Curvilinear Projection [pdf] - Gaofeng Meng, Zuming Huang, Yonghong Song, Shiming Xiang, Chunhong Pan
  • Predicting Deep Zero-Shot Convolutional Neural Networks Using Textual Descriptions [pdf] - Jimmy Lei Ba, Kevin Swersky, Sanja Fidler, Ruslan salakhutdinov
  • Lost Shopping! Monocular Localization in Large Indoor Spaces [pdf] - Shenlong Wang, Sanja Fidler, Raquel Urtasun
  • FASText: Efficient Unconstrained Scene Text Detector [pdf] - Michal Busta, Lukas Neumann, Jiri Matas
The Joint Image Handbook [pdf]
Matthew Trager, Martial Hebert, Jean Ponce

Abstract: Given multiple perspective photographs, point correspondences form the joint image, effectively a replica of three-dimensional space distributed across its two-dimensional projections. This set can be characterized by multilinear equations over image coordinates, such as epipolar and trifocal constraints. We revisit in this paper the geometric and algebraic properties of the joint image, and address fundamental questions such as how many and which multilinearities are necessary and/or sufficient to determine camera geometry and/or image correspondences. The new theoretical results in this paper answer these questions in a very general setting and, in turn, are intended to serve as a handbook reference about multilinearities for practitioners.
Similar papers:
  • Joint Camera Clustering and Surface Segmentation for Large-Scale Multi-View Stereo [pdf] - Runze Zhang, Shiwei Li, Tian Fang, Siyu Zhu, Long Quan
  • Constrained Convolutional Neural Networks for Weakly Supervised Segmentation [pdf] - Deepak Pathak, Philipp Krahenbuhl, Trevor Darrell
  • Robust and Optimal Sum-of-Squares-Based Point-to-Plane Registration of Image Sets and Structured Scenes [pdf] - Danda Pani Paudel, Adlane Habed, Cedric Demonceaux, Pascal Vasseur
  • Structure From Motion Using Structure-Less Resection [pdf] - Enliang Zheng, Changchang Wu
  • Wide Baseline Stereo Matching With Convex Bounded Distortion Constraints [pdf] - Meirav Galun, Tal Amir, Tal Hassner, Ronen Basri, Yaron Lipman
Learning Spatiotemporal Features With 3D Convolutional Networks [pdf]
Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, Manohar Paluri

Abstract: We propose a simple, yet effective approach for spatiotemporal feature learning using deep 3-dimensional convolutional networks (3D ConvNets) trained on a large scale supervised video dataset. Our findings are three-fold: 1) 3D ConvNets are more suitable for spatiotemporal feature learning compared to 2D ConvNets; 2) A homogeneous architecture with small 3 3 3 convolution kernels in all layers is among the best performing architectures for 3D ConvNets; and 3) Our learned features, namely C3D (Convolutional 3D), with a simple linear classifier outperform state-of-the-art methods on 4 different benchmarks and are comparable with current best methods on the other 2 benchmarks. In addition, the features are compact: achieving 52.8% accuracy on UCF101 dataset with only 10 dimensions and also very efficient to compute due to the fast inference of ConvNets. Finally, they are conceptually very simple and easy to train and use.
Similar papers:
  • Learning to Track for Spatio-Temporal Action Localization [pdf] - Philippe Weinzaepfel, Zaid Harchaoui, Cordelia Schmid
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Action Recognition by Hierarchical Mid-Level Action Elements [pdf] - Tian Lan, Yuke Zhu, Amir Roshan Zamir, Silvio Savarese
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
  • Human Action Recognition Using Factorized Spatio-Temporal Convolutional Networks [pdf] - Lin Sun, Kui Jia, Dit-Yan Yeung, Bertram E. Shi
Pose Induction for Novel Object Categories [pdf]
Shubham Tulsiani, Joao Carreira, Jitendra Malik

Abstract: We address the task of predicting pose for objects of unannotated object categories from a small seed set of annotated object classes. We present a generalized classifier that can reliably induce pose given a single instance of a novel category. In case of availability of a large collection of novel instances, our approach then jointly reasons over all instances to improve the initial estimates. We empirically validate the various components of our algorithm and quantitatively show that our method produces reliable pose estimates. We also show qualitative results on a diverse set of classes and further demonstrate the applicability of our system for learning shape models of novel object classes.
Similar papers:
  • Semantic Pose Using Deep Networks Trained on Synthetic RGB-D [pdf] - Jeremie Papon, Markus Schoeler
  • Learning Analysis-by-Synthesis for 6D Pose Estimation in RGB-D Images [pdf] - Alexander Krull, Eric Brachmann, Frank Michel, Michael Ying Yang, Stefan Gumhold, Carsten Rother
  • Real-Time Pose Estimation Piggybacked on Object Detection [pdf] - Roman Juranek, Adam Herout, Marketa Dubska, Pavel Zemcik
  • Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf] - Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang
  • Maximum-Margin Structured Learning With Deep Networks for 3D Human Pose Estimation [pdf] - Sijin Li, Weichen Zhang, Antoni B. Chan
Regressing a 3D Face Shape From a Single Image [pdf]
Sergey Tulyakov, Nicu Sebe

Abstract: In this work we present a method to estimate a 3D face shape from a single image. Our method is based on a cascade regression framework that directly estimates face landmarks locations in 3D. We include the knowledge that a face is a 3D object into the learning pipeline and show how this information decreases localization errors while keeping the computational time low. We predict the actual positions of the landmarks even if they are occluded due to face rotation. To support the ability of our method to reliably reconstruct 3D shapes, we introduce a simple method for head pose estimation using a single image that reaches higher accuracy than the state of the art. Comparison of 3D face landmarks localization with the available state of the art further supports the feasibility of a single-step face shape estimation. The code, trained models and our 3D annotations will be made available to the research community.
Similar papers:
  • Robust Model-Based 3D Head Pose Estimation [pdf] - Gregory P. Meyer, Shalini Gupta, Iuri Frosio, Dikpal Reddy, Jan Kautz
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Regressive Tree Structured Model for Facial Landmark Localization [pdf] - Gee-Sern Hsu, Kai-Hsiang Chang, Shih-Chieh Huang
  • Leveraging Datasets With Varying Annotations for Face Alignment via Deep Regression Network [pdf] - Jie Zhang, Meina Kan, Shiguang Shan, Xilin Chen
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
Simultaneous Deep Transfer Across Domains and Tasks [pdf]
Eric Tzeng, Judy Hoffman, Trevor Darrell, Kate Saenko

Abstract: Source domain! Recent reports suggest that a generic supervised deep CNN model trained on a large-scale dataset reduces, but does not remove, dataset bias. Fine-tuning deep models in a new domain can require a significant amount of labeled data, which for many applications is simply not available. We propose a new CNN architecture to exploit unlabeled and sparsely labeled target domain data. Our approach simultaneously optimizes for domain invariance to facilitate domain transfer and uses a soft label distribution matching loss to transfer information between tasks. Our proposed adaptation method offers empirical performance which exceeds previously published results on two standard benchmark visual domain adaptation tasks, evaluated across supervised and semi-supervised adaptation settings.
Similar papers:
  • Zero-Shot Learning via Semantic Similarity Embedding [pdf] - Ziming Zhang, Venkatesh Saligrama
  • Multi-View Domain Generalization for Visual Recognition [pdf] - Li Niu, Wen Li, Dong Xu
  • Unsupervised Domain Adaptation for Zero-Shot Learning [pdf] - Elyor Kodirov, Tao Xiang, Zhenyong Fu, Shaogang Gong
  • Unsupervised Domain Adaptation With Imbalanced Cross-Domain Data [pdf] - Tzu Ming Harry Hsu, Wei Yu Chen, Cheng-An Hou, Yao-Hung Hubert Tsai, Yi-Ren Yeh, Yu-Chiang Frank Wang
  • Bi-Shifting Auto-Encoder for Unsupervised Domain Adaptation [pdf] - Meina Kan, Shiguang Shan, Xilin Chen
3D Object Reconstruction From Hand-Object Interactions [pdf]
Dimitrios Tzionas, Juergen Gall

Abstract: Recent advances have enabled 3d object reconstruction approaches using a single off-the-shelf RGB-D camera. Although these approaches are successful for a wide range of object classes, they rely on stable and distinctive geometric or texture features. Many objects like mechanical parts, toys, household or decorative articles, however, are textureless and characterized by minimalistic shapes that are simple and symmetric. Existing in-hand scanning systems and 3d reconstruction techniques fail for such symmetric objects in the absence of highly distinctive features. In this work, we show that extracting 3d hand motion for in-hand scanning effectively facilitates the reconstruction of even featureless and highly symmetric objects and we present an approach that fuses the rich additional information of hands into a 3d reconstruction pipeline, significantly contributing to the state-of-the-art of in-hand scanning. Figure 1. Reconstruction of a symmetric, textureless object. Both the front and the bottom view are provided for better visualization. Left: Existing in-hand scanning approaches fail for such objects. Middle and right: Successful reconstruction by the proposed inhand scanning system that incorporates 3d hand motion capture.
Similar papers:
  • Lending A Hand: Detecting Hands and Recognizing Activities in Complex Egocentric Interactions [pdf] - Sven Bambach, Stefan Lee, David J. Crandall, Chen Yu
  • 3D Hand Pose Estimation Using Randomized Decision Forest With Segmentation Index Points [pdf] - Peiyi Li, Haibin Ling, Xi Li, Chunyuan Liao
  • Understanding Everyday Hands in Action From RGB-D Images [pdf] - Gregory Rogez, James S. Supancic III, Deva Ramanan
  • A Collaborative Filtering Approach to Real-Time Hand Pose Estimation [pdf] - Chiho Choi, Ayan Sinha, Joon Hee Choi, Sujin Jang, Karthik Ramani
  • Depth-Based Hand Pose Estimation: Data, Methods, and Challenges [pdf] - James S. Supancic III, Gregory Rogez, Yi Yang, Jamie Shotton, Deva Ramanan
Global, Dense Multiscale Reconstruction for a Billion Points [pdf]
Benjamin Ummenhofer, Thomas Brox

Abstract: We present a variational approach for surface reconstruction from a set of oriented points with scale information. We focus particularly on scenarios with non-uniform point densities due to images taken from different distances. In contrast to previous methods, we integrate the scale information in the objective and globally optimize the signed distance function of the surface on a balanced octree grid. We use a finite element discretization on the dual structure of the octree minimizing the number of variables. The tetrahedral mesh is generated efficiently from the dual structure, and also memory efficiency is optimized, such that robust data terms can be used even on very large scenes. The surface normals are explicitly optimized and used for surface extraction to improve the reconstruction at edges and corners.
Similar papers:
  • Dense Image Registration and Deformable Surface Reconstruction in Presence of Occlusions and Minimal Texture [pdf] - Dat Tien Ngo, Sanghyuk Park, Anne Jorstad, Alberto Crivellaro, Chang D. Yoo, Pascal Fua
  • Exploiting High Level Scene Cues in Stereo Reconstruction [pdf] - Simon Hadfield, Richard Bowden
  • Peeking Template Matching for Depth Extension [pdf] - Simon Korman, Eyal Ofek, Shai Avidan
  • Variational PatchMatch MultiView Reconstruction and Refinement [pdf] - Philipp Heise, Brian Jensen, Sebastian Klose, Alois Knoll
  • Exploiting Object Similarity in 3D Reconstruction [pdf] - Chen Zhou, Fatma Guney, Yizhou Wang, Andreas Geiger
Common Subspace for Model and Similarity: Phrase Learning for Caption Generation From Images [pdf]
Yoshitaka Ushiku, Masataka Yamaguchi, Yusuke Mukuta, Tatsuya Harada

Abstract: Generating captions to describe images is a fundamental problem that combines computer vision and natural language processing. Recent works focus on descriptive phrases, such as a white dog to explain the visual composites of an input image. The phrases can not only express objects, attributes, events, and their relations but can also reduce visual complexity. A caption for an input image can be generated by connecting estimated phrases using a grammar model. However, because phrases are combinations of various words, the number of phrases is much larger than the number of single words. Consequently, the accuracy of phrase estimation suffers from too few training samples per phrase. In this paper, we propose a novel phrase-learning method: Common Subspace for Model and Similarity (CoSMoS). In order to overcome the shortage of training samples, CoSMoS obtains a subspace in which (a) all feature vectors associated with the same phrase are mapped as mutually close, (b) classifiers for each phrase are learned, and (c) training samples are shared among co-occurring phrases. Experimental results demonstrate that our system is more accurate than those in earlier work and that the accuracy increases when the dataset from the web increases.
Similar papers:
  • Segment-Phrase Table for Semantic Segmentation, Visual Entailment and Paraphrasing [pdf] - Hamid Izadinia, Fereshteh Sadeghi, Santosh K. Divvala, Hannaneh Hajishirzi, Yejin Choi, Ali Farhadi
  • Visual Phrases for Exemplar Face Detection [pdf] - Vijay Kumar, Anoop Namboodiri, C. V. Jawahar
  • Multimodal Convolutional Neural Networks for Matching Image and Sentence [pdf] - Lin Ma, Zhengdong Lu, Lifeng Shang, Hang Li
  • Guiding the Long-Short Term Memory Model for Image Caption Generation [pdf] - Xu Jia, Efstratios Gavves, Basura Fernando, Tinne Tuytelaars
  • Flickr30k Entities: Collecting Region-to-Phrase Correspondences for Richer Image-to-Sentence Models [pdf] - Bryan A. Plummer, Liwei Wang, Chris M. Cervantes, Juan C. Caicedo, Julia Hockenmaier, Svetlana Lazebnik
Multiresolution Hierarchy Co-Clustering for Semantic Segmentation in Sequences With Small Variations [pdf]
David Varas, Monica Alfaro, Ferran Marques

Abstract: This paper presents a co-clustering technique that, given a collection of images and their hierarchies, clusters nodes from these hierarchies to obtain a coherent multiresolution representation of the image collection. We formalize the co-clustering as a Quadratic Semi-Assignment Problem and solve it with a linear programming relaxation approach that makes effective use of information from hierarchies. Initially, we address the problem of generating an optimal, coherent partition per image and, afterwards, we extend this method to a multiresolution framework. Finally, we particularize this framework to an iterative multiresolution video segmentation algorithm in sequences with small variations. We evaluate the algorithm on the Video Occlusion/Object Boundary Detection Dataset, showing that it produces state-of-the-art results in these scenarios.
Similar papers:
  • Self-Occlusions and Disocclusions in Causal Video Object Segmentation [pdf] - Yanchao Yang, Ganesh Sundaramoorthi, Stefano Soatto
  • Contour Detection and Characterization for Asynchronous Event Sensors [pdf] - Francisco Barranco, Ching L. Teo, Cornelia Fermuller, Yiannis Aloimonos
  • Efficient Video Segmentation Using Parametric Graph Partitioning [pdf] - Chen-Ping Yu, Hieu Le, Gregory Zelinsky, Dimitris Samaras
  • High-for-Low and Low-for-High: Efficient Boundary Detection From Deep Object Features and its Applications to High-Level Vision [pdf] - Gedas Bertasius, Jianbo Shi, Lorenzo Torresani
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
Learning Common Sense Through Visual Abstraction [pdf]
Ramakrishna Vedantam, Xiao Lin, Tanmay Batra, C. Lawrence Zitnick, Devi Parikh

Abstract: on Ramakrishna Vedantam1 Xiao Lin1 Tanmay Batra2 C. Lawrence Zitnick3 Devi Parikh1 1 Virginia Tech 2 Carnegie Mellon University 3 Microsoft Research 1 {vrama91,linxiao,parikh}@vt.edu 2 tbatra@andrew.cmu.edu
Similar papers:
  • Weakly-Supervised Alignment of Video With Text [pdf] - Piotr Bojanowski, Remi Lajugie, Edouard Grave, Francis Bach, Ivan Laptev, Jean Ponce, Cordelia Schmid
  • Lost Shopping! Monocular Localization in Large Indoor Spaces [pdf] - Shenlong Wang, Sanja Fidler, Raquel Urtasun
  • FASText: Efficient Unconstrained Scene Text Detector [pdf] - Michal Busta, Lukas Neumann, Jiri Matas
  • Predicting Deep Zero-Shot Convolutional Neural Networks Using Textual Descriptions [pdf] - Jimmy Lei Ba, Kevin Swersky, Sanja Fidler, Ruslan salakhutdinov
  • Text Flow: A Unified Text Detection System in Natural Scene Images [pdf] - Shangxuan Tian, Yifeng Pan, Chang Huang, Shijian Lu, Kai Yu, Chew Lim Tan
Differential Recurrent Neural Networks for Action Recognition [pdf]
Vivek Veeriah, Naifan Zhuang, Guo-Jun Qi

Abstract: The long short-term memory (LSTM) neural network is capable of processing complex sequential information since it utilizes special gating schemes for learning representations from long input sequences. It has the potential to model any sequential time-series data, where the current hidden state has to be considered in the context of the past hidden states. This property makes LSTM an ideal choice to learn the complex dynamics of various actions. Unfortunately, the conventional LSTMs do not consider the impact of spatio-temporal dynamics corresponding to the given salient motion patterns, when they gate the information that ought to be memorized through time. To address this problem, we propose a differential gating scheme for the LSTM neural network, which emphasizes on the change in information gain caused by the salient motions between the successive frames. This change in information gain is quantied by Derivative of States (DoS), and thus the proposed LSTM model is termed as differential Recurrent Neural Network (dRNN). We demonstrate the effectiveness of the proposed model by automatically recognizing actions from the real-world 2D and 3D human action datasets. Our study is one of the rst works towards demonstrating the potential of learning complex time-series representations via high-order derivatives of states.
Similar papers:
  • Action Localization in Videos Through Context Walk [pdf] - Khurram Soomro, Haroon Idrees, Mubarak Shah
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
  • Category-Blind Human Action Recognition: A Practical Recognition System [pdf] - Wenbo Li, Longyin Wen, Mooi Choo Chuah, Siwei Lyu
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
Learning Visual Clothing Style With Heterogeneous Dyadic Co-Occurrences [pdf]
Andreas Veit, Balazs Kovacs, Sean Bell, Julian McAuley, Kavita Bala, Serge Belongie

Abstract: With the rapid proliferation of smart mobile devices, users now take millions of photos every day. These include large numbers of clothing and accessory images. We would like to answer questions like What outfit goes well with this pair of shoes? To answer these types of questions, one has to go beyond learning visual similarity and learn a visual notion of compatibility across categories. In this paper, we propose a novel learning framework to help answer these types of questions. The main idea of this framework is to learn a feature transformation from images of items into a latent space that expresses compatibility. For the feature transformation, we use a Siamese Convolutional Neural Network (CNN) architecture, where training examples are pairs of items that are either compatible or incompatible. We model compatibility based on co-occurrence in largescale user behavior data; in particular co-purchase data from Amazon.com. To learn cross-category fit, we introduce a strategic method to sample training data, where pairs of items are heterogeneous dyads, i.e., the two elements of a pair belong to different high-level categories. While this approach is applicable to a wide variety of settings, we focus on the representative problem of learning compatible clothing style. Our results indicate that the proposed framework is capable of learning semantic information about visual style and is able to generate outfits of clothes, with items from different categories, that go well
Similar papers:
  • Im2Calories: Towards an Automated Mobile Vision Food Diary [pdf] - Austin Meyers, Nick Johnston, Vivek Rathod, Anoop Korattikara, Alex Gorban, Nathan Silberman, Sergio Guadarrama, George Papandreou, Jonathan Huang, Kevin P. Murphy
  • Understanding Deep Features With Computer-Generated Imagery [pdf] - Mathieu Aubry, Bryan C. Russell
  • Cross-Domain Image Retrieval With a Dual Attribute-Aware Ranking Network [pdf] - Junshi Huang, Rogerio S. Feris, Qiang Chen, Shuicheng Yan
  • Single Image 3D Without a Single 3D Image [pdf] - David F. Fouhey, Wajahat Hussain, Abhinav Gupta, Martial Hebert
  • Where to Buy It: Matching Street Clothing Photos in Online Shops [pdf] - M. Hadi Kiapour, Xufeng Han, Svetlana Lazebnik, Alexander C. Berg, Tamara L. Berg
Unsupervised Cross-Modal Synthesis of Subject-Specific Scans [pdf]
Raviteja Vemulapalli, Hien Van Nguyen, Shaohua Kevin Zhou

Abstract: Recently, cross-modal synthesis of subject-specific scans has been receiving significant attention from the medical imaging community. Though various synthesis approaches have been introduced in the recent past, most of them are either tailored to a specific application or proposed for the supervised setting, i.e., they assume the availability of training data from the same set of subjects in both source and target modalities. But, collecting multiple scans from each subject is undesirable. Hence, to address this issue, we propose a general unsupervised cross-modal medical image synthesis approach that works without paired training data. Given a source modality image of a subject, we first generate multiple target modality candidate values for each voxel independently using cross-modal nearest neighbor search. Then, we select the best candidate values jointly for all the voxels by simultaneously maximizing a global mutual information cost function and a local spatial consistency cost function. Finally, we use coupled sparse representation for further refinement of synthesized images. Our experiments on generating T1-MRI brain scans from T2-MRI and vice versa demonstrate that the synthesis capability of the proposed unsupervised approach is comparable to various state-of-the-art supervised approaches in the literature.
Similar papers:
  • Unsupervised Domain Adaptation for Zero-Shot Learning [pdf] - Elyor Kodirov, Tao Xiang, Zhenyong Fu, Shaogang Gong
  • 3D-Assisted Feature Synthesis for Novel Views of an Object [pdf] - Hao Su, Fan Wang, Eric Yi, Leonidas J. Guibas
  • MMSS: Multi-Modal Sharable and Specific Feature Learning for RGB-D Object Recognition [pdf] - Anran Wang, Jianfei Cai, Jiwen Lu, Tat-Jen Cham
  • Bi-Shifting Auto-Encoder for Unsupervised Domain Adaptation [pdf] - Meina Kan, Shiguang Shan, Xilin Chen
  • Conditional Convolutional Neural Network for Modality-Aware Face Recognition [pdf] - Chao Xiong, Xiaowei Zhao, Danhang Tang, Karlekar Jayashree, Shuicheng Yan, Tae-Kyun Kim
Simpler Non-Parametric Methods Provide as Good or Better Results to Multiple-Instance Learning [pdf]
Ragav Venkatesan, Parag Chandakkar, Baoxin Li

Abstract: Multiple-instance learning (MIL) is a unique learning problem in which training data labels are available only for collections of objects (called bags) instead of individual objects (called instances). A plethora of approaches have been developed to solve this problem in the past years. Popular methods include the diverse density, MILIS and DD-SVM. While having been widely used, these methods, particularly those in computer vision have attempted fairly sophisticated solutions to solve certain unique and particular configurations of the MIL space. In this paper, we analyze the MIL feature space using modified versions of traditional non-parametric techniques like the Parzen window and k-nearest-neighbour, and develop a learning approach employing distances to k-nearest neighbours of a point in the feature space. We show that these methods work as well, if not better than most recently published methods on benchmark datasets. We compare and contrast our analysis with the well-established diversedensity approach and its variants in recent literature, using benchmark datasets including the Musk, Andrews and Corel datasets, along with a diabetic retinopathy pathology diagnosis dataset. Experimental results demonstrate that, while enjoying an intuitive interpretation and supporting fast learning, these method have the potential of delivering improved performance even for complex data arising from real-world applications.
Similar papers:
  • AttentionNet: Aggregating Weak Directions for Accurate Object Detection [pdf] - Donggeun Yoo, Sunggyun Park, Joon-Young Lee, Anthony S. Paek, In So Kweon
  • Monocular Object Instance Segmentation and Depth Ordering With CNNs [pdf] - Ziyu Zhang, Alexander G. Schwing, Sanja Fidler, Raquel Urtasun
  • A Self-Paced Multiple-Instance Learning Framework for Co-Saliency Detection [pdf] - Dingwen Zhang, Deyu Meng, Chao Li, Lu Jiang, Qian Zhao, Junwei Han
  • Towards Computational Baby Learning: A Weakly-Supervised Approach for Object Detection [pdf] - Xiaodan Liang, Si Liu, Yunchao Wei, Luoqi Liu, Liang Lin, Shuicheng Yan
  • Relaxed Multiple-Instance SVM With Application to Object Discovery [pdf] - Xinggang Wang, Zhuotun Zhu, Cong Yao, Xiang Bai
An Efficient Minimal Solution for Multi-Camera Motion [pdf]
Jonathan Ventura, Clemens Arth, Vincent Lepetit

Abstract: We propose an efficient method for estimating the motion of a multi-camera rig from a minimal set of feature correspondences. Existing methods for solving the multi-camera relative pose problem require extra correspondences, are slow to compute, and/or produce a multitude of solutions. Our solution uses a first-order approximation to relative pose in order to simplify the problem and produce an accurate estimate quickly. The solver is applicable to sequential multi-camera motion estimation and is fast enough for real-time implementation in a random sampling framework. Our experiments show that our approach is both stable and efficient on challenging test sequences.
Similar papers:
  • Minimal Solvers for 3D Geometry From Satellite Imagery [pdf] - Enliang Zheng, Ke Wang, Enrique Dunn, Jan-Michael Frahm
  • Global Structure-From-Motion by Similarity Averaging [pdf] - Zhaopeng Cui, Ping Tan
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
  • Efficient Solution to the Epipolar Geometry for Radially Distorted Cameras [pdf] - Zuzana Kukelova, Jan Heller, Martin Bujnak, Andrew Fitzgibbon, Tomas Pajdla
  • Structure From Motion Using Structure-Less Resection [pdf] - Enliang Zheng, Changchang Wu
Sequence to Sequence - Video to Text [pdf]
Subhashini Venugopalan, Marcus Rohrbach, Jeffrey Donahue, Raymond Mooney, Trevor Darrell, Kate Saenko

Abstract: Raw Frames Real-world videos often have complex dynamics; and methods for generating open-domain video descriptions should be sensitive to temporal structure and allow both input (sequence of frames) and output (sequence of words) of variable length. To approach this problem, we propose a novel end-to-end sequence-to-sequence model to generate captions for videos. For this we exploit recurrent neural networks, specifically LSTMs, which have demonstrated stateof-the-art performance in image caption generation. Our LSTM model is trained on video-sentence pairs and learns to associate a sequence of video frames to a sequence of words in order to generate a description of the event in the video clip. Our model naturally is able to learn the temporal structure of the sequence of frames as well as the sequence model of the generated sentences, i.e. a language model. We evaluate several variants of our model that exploit different visual features on a standard set of YouTube videos and two movie description datasets (M-VAD and MPII-MD).
Similar papers:
  • Unsupervised Extraction of Video Highlights Via Robust Recurrent Auto-Encoders [pdf] - Huan Yang, Baoyuan Wang, Stephen Lin, David Wipf, Minyi Guo, Baining Guo
  • Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books [pdf] - Yukun Zhu, Ryan Kiros, Rich Zemel, Ruslan Salakhutdinov, Raquel Urtasun, Antonio Torralba, Sanja Fidler
  • Differential Recurrent Neural Networks for Action Recognition [pdf] - Vivek Veeriah, Naifan Zhuang, Guo-Jun Qi
  • Guiding the Long-Short Term Memory Model for Image Caption Generation [pdf] - Xu Jia, Efstratios Gavves, Basura Fernando, Tinne Tuytelaars
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
Leave-One-Out Kernel Optimization for Shadow Detection [pdf]
Tomas F. Yago Vicente, Minh Hoai, Dimitris Samaras

Abstract: The objective of this work is to detect shadows in images. We pose this as the problem of labeling image regions, where each region corresponds to a group of superpixels. To predict the label of each region, we train a kernel Least-Squares SVM for separating shadow and non-shadow regions. The parameters of the kernel and the classifier are jointly learned to minimize the leave-one-out cross validation error. Optimizing the leave-one-out cross validation error is typically difficult, but it can be done efficiently in our framework. Experiments on two challenging shadow datasets, UCF and UIUC, show that our region classifier outperforms more complex methods. We further enhance the performance of the region classifier by embedding it in an MRF framework and adding pairwise contextual cues. This leads to a method that significantly outperforms the state-of-the-art.
Similar papers:
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
  • Secrets of GrabCut and Kernel K-Means [pdf] - Meng Tang, Ismail Ben Ayed, Dmitrii Marin, Yuri Boykov
  • Per-Sample Kernel Adaptation for Visual Recognition and Grouping [pdf] - Borislav Antic, Bjorn Ommer
  • An Accurate Iris Segmentation Framework Under Relaxed Imaging Constraints Using Total Variation Model [pdf] - Zijing Zhao, Kumar Ajay
  • Illumination Robust Color Naming via Label Propagation [pdf] - Yuanliu liu, Zejian Yuan, Badong Chen, Jianru Xue, Nanning Zheng
Context-Aware CNNs for Person Head Detection [pdf]
Tuan-Hung Vu, Anton Osokin, Ivan Laptev

Abstract: Person detection is a key problem for many computer vision tasks. While face detection has reached maturity, detecting people under a full variation of camera view-points, human poses, lighting conditions and occlusions is still a difficult challenge. In this work we focus on detecting human heads in natural scenes. Starting from the recent local R-CNN object detector, we extend it with two types of contextual cues. First, we leverage person-scene relations and propose a Global CNN model trained to predict positions and scales of heads directly from the full image. Second, we explicitly model pairwise relations among objects and train a Pairwise CNN model using a structured-output surrogate loss. The Local, Global and Pairwise models are combined into a joint CNN framework. To train and test our full model, we introduce a large dataset composed of 369, 846 human heads annotated in 224, 740 movie frames. We evaluate our method and demonstrate improvements of person head detection against several recent baselines in three datasets. We also show improvements of the detection speed provided by our model.
Similar papers:
  • Person Recognition in Personal Photo Collections [pdf] - Seong Joon Oh, Rodrigo Benenson, Mario Fritz, Bernt Schiele
  • Conditional Random Fields as Recurrent Neural Networks [pdf] - Shuai Zheng, Sadeep Jayasumana, Bernardino Romera-Paredes, Vibhav Vineet, Zhizhong Su, Dalong Du, Chang Huang, Philip H. S. Torr
  • Joint Object and Part Segmentation Using Deep Learned Potentials [pdf] - Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille
  • Semantic Image Segmentation via Deep Parsing Network [pdf] - Ziwei Liu, Xiaoxiao Li, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
Dense Optical Flow Prediction From a Static Image [pdf]
Jacob Walker, Abhinav Gupta, Martial Hebert

Abstract: Given a scene, what is going to move, and in what direction will it move? Such a question could be considered a non-semantic form of action prediction. In this work, we present a convolutional neural network (CNN) based approach for motion prediction. Given a static image, this CNN predicts the future motion of each and every pixel in the image in terms of optical flow. Our CNN model leverages the data in tens of thousands of realistic videos to train our model. Our method relies on absolutely no human labeling and is able to predict motion based on the context of the scene. Because our CNN model makes no assumptions about the underlying scene, it can predict future optical flow on a diverse set of scenarios. We outperform all previous approaches by large margins.
Similar papers:
  • Flowing ConvNets for Human Pose Estimation in Videos [pdf] - Tomas Pfister, James Charles, Andrew Zisserman
  • Face Flow [pdf] - Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou
  • Unsupervised Tube Extraction Using Transductive Learning and Dense Trajectories [pdf] - Mihai Marian Puscas, Enver Sangineto, Dubravko Culibrk, Nicu Sebe
  • Recurrent Network Models for Human Dynamics [pdf] - Katerina Fragkiadaki, Sergey Levine, Panna Felsen, Jitendra Malik
  • FlowNet: Learning Optical Flow With Convolutional Networks [pdf] - Alexey Dosovitskiy, Philipp Fischer, Eddy Ilg, Philip Hausser, Caner Hazirbas, Vladimir Golkov, Patrick van der Smagt, Daniel Cremers, Thomas Brox
Adaptively Unified Semi-Supervised Dictionary Learning With Active Points [pdf]
Xiaobo Wang, Xiaojie Guo, Stan Z. Li

Abstract: Semi-supervised dictionary learning aims to construct a dictionary by utilizing both labeled and unlabeled data. To enhance the discriminative capability of the learned dictionary, numerous discriminative terms have been proposed by evaluating either the prediction loss or the class separation criterion on the coding vectors of labeled data, but with rare consideration of the power of the coding vectors corresponding to unlabeled data. In this paper, we present a novel semi-supervised dictionary learning method, which uses the informative coding vectors of both labeled and unlabeled data, and adaptively emphasizes the high confidence coding vectors of unlabeled data to enhance the dictionary discriminative capability simultaneously. By doing so, we integrate the discrimination of dictionary, the induction of classifier to new testing data and the transduction of labels to unlabeled data into a unified framework. To solve the proposed problem, an effective iterative algorithm is designed. Experimental results on a series of benchmark databases show that our method outperforms other stateof-the-art dictionary learning methods in most cases.
Similar papers:
  • Person Re-Identification With Discriminatively Trained Viewpoint Invariant Dictionaries [pdf] - Srikrishna Karanam, Yang Li, Richard J. Radke
  • Convolutional Sparse Coding for Image Super-Resolution [pdf] - Shuhang Gu, Wangmeng Zuo, Qi Xie, Deyu Meng, Xiangchu Feng, Lei Zhang
  • Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf] - Yuhui Quan, Yan Huang, Hui Ji
  • Conformal and Low-Rank Sparse Representation for Image Restoration [pdf] - Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng
  • Linearization to Nonlinear Learning for Visual Tracking [pdf] - Bo Ma, Hongwei Hu, Jianbing Shen, Yuping Zhang, Fatih Porikli
Additive Nearest Neighbor Feature Maps [pdf]
Zhenzhen Wang, Xiao-Tong Yuan, Qingshan Liu, Shuicheng Yan

Abstract: In this paper, we present a concise framework to approximately construct feature maps for nonlinear additive kernels such as the Intersection, Hellingers, and 2 kernels. The core idea is to construct a set of anchor points for each individual feature and assign to every query the feature map of its nearest neighbor or the weighted combination of those of its k-nearest neighbors in the anchors. The constructed feature maps can be compactly stored by a group of nearest neighbor (binary) indication vectors along with the anchor feature maps. The approximation error of such an anchored feature mapping approach is analyzed. We evaluate the performance of our approach on large-scale nonlinear support vector machines (SVMs) learning tasks in the context of visual object classification. Experimental results on several benchmark data sets show the superiority of our method over existing feature mapping methods in achieving reasonable tradeoff between training time and testing accuracy.
Similar papers:
  • Secrets of GrabCut and Kernel K-Means [pdf] - Meng Tang, Ismail Ben Ayed, Dmitrii Marin, Yuri Boykov
  • Beyond Covariance: Feature Representation With Nonlinear Kernel Matrices [pdf] - Lei Wang, Jianjia Zhang, Luping Zhou, Chang Tang, Wanqing Li
  • Scalable Nonlinear Embeddings for Semantic Category-Based Image Retrieval [pdf] - Gaurav Sharma, Bernt Schiele
  • Per-Sample Kernel Adaptation for Visual Recognition and Grouping [pdf] - Borislav Antic, Bjorn Ommer
  • Low Dimensional Explicit Feature Maps [pdf] - Ondrej Chum
Beyond Covariance: Feature Representation With Nonlinear Kernel Matrices [pdf]
Lei Wang, Jianjia Zhang, Luping Zhou, Chang Tang, Wanqing Li

Abstract: Covariance matrix has recently received increasing attention in computer vision by leveraging Riemannian geometry of symmetric positive-definite (SPD) matrices. Originally proposed as a region descriptor, it has now been used as a generic representation in various recognition tasks. However, covariance matrix has shortcomings such as being prone to be singular, limited capability in modeling complicated feature relationship, and having a fixed form of representation. This paper argues that more appropriate SPD-matrix-based representations shall be explored to achieve better recognition. It proposes an open framework to use the kernel matrix over feature dimensions as a generic representation and discusses its properties and advantages. The proposed framework significantly elevates covariance representation to the unlimited opportunities provided by this new representation. Experimental study shows that this representation consistently outperforms its covariance counterpart on various visual recognition tasks. In particular, it achieves significant improvement on skeleton-based human action recognition, demonstrating the state-of-the-art performance over both the covariance and the existing non-covariance representations.
Similar papers:
  • Additive Nearest Neighbor Feature Maps [pdf] - Zhenzhen Wang, Xiao-Tong Yuan, Qingshan Liu, Shuicheng Yan
  • Secrets of GrabCut and Kernel K-Means [pdf] - Meng Tang, Ismail Ben Ayed, Dmitrii Marin, Yuri Boykov
  • Beyond Gauss: Image-Set Matching on the Riemannian Manifold of PDFs [pdf] - Mehrtash Harandi, Mathieu Salzmann, Mahsa Baktashmotlagh
  • Low Dimensional Explicit Feature Maps [pdf] - Ondrej Chum
  • Per-Sample Kernel Adaptation for Visual Recognition and Grouping [pdf] - Borislav Antic, Bjorn Ommer
BodyPrint: Pose Invariant 3D Shape Matching of Human Bodies [pdf]
Jiangping Wang, Kai Ma, Vivek Kumar Singh, Thomas Huang, Terrence Chen

Abstract: 3D human body shape matching has substantial potential in many real world applications, especially with recent advances in 3D range sensing technology. We address this problem by proposing a novel holistic human body shape descriptor called BodyPrint. To compute the bodyprint for a given body scan, we fit a deformable human body mesh, and project the mesh parameters to a low-dimensional subspace which improves discriminability across different persons. Experiments are carried out on three real-world human body datasets to demonstrate that BodyPrint is robust to pose variation as well as missing information and sensor noise. It improves the matching accuracy significantly compared to conventional 3D shape matching techniques using local features. To facilitate practical applications where the shape database may grow over time, we also extend our learning framework to handle online updates.
Similar papers:
  • The One Triangle Three Parallelograms Sampling Strategy and Its Application in Shape Regression [pdf] - Mikael Nilsson
  • Human Pose Estimation in Videos [pdf] - Dong Zhang, Mubarak Shah
  • Multi-View Convolutional Neural Networks for 3D Shape Recognition [pdf] - Hang Su, Subhransu Maji, Evangelos Kalogerakis, Erik Learned-Miller
  • Compositional Hierarchical Representation of Shape Manifolds for Classification of Non-Manifold Shapes [pdf] - Mete Ozay, Umit Rusen Aktas, Jeremy L. Wyatt, Ales Leonardis
  • Detailed Full-Body Reconstructions of Moving People From Monocular RGB-D Sequences [pdf] - Federica Bogo, Michael J. Black, Matthew Loper, Javier Romero
Deep Networks for Image Super-Resolution With Sparse Prior [pdf]
Zhaowen Wang, Ding Liu, Jianchao Yang, Wei Han, Thomas Huang

Abstract: Deep learning techniques have been successfully applied in many areas of computer vision, including low-level image restoration problems. For image super-resolution, several models based on deep neural networks have been recently proposed and attained superior performance that overshadows all previous handcrafted models. The question then arises whether large-capacity and data-driven models have become the dominant solution to the ill-posed superresolution problem. In this paper, we argue that domain expertise represented by the conventional sparse coding model is still valuable, and it can be combined with the key ingredients of deep learning to achieve further improved results. We show that a sparse coding model particularly designed for super-resolution can be incarnated as a neural network, and trained in a cascaded structure from end to end. The interpretation of the network based on sparse coding leads to much more efficient and effective training, as well as a reduced model size. Our model is evaluated on a wide range of images, and shows clear advantage over existing state-of-the-art methods in terms of both restoration accuracy and human subjective quality.
Similar papers:
  • Removing Rain From a Single Image via Discriminative Sparse Coding [pdf] - Yu Luo, Yong Xu, Hui Ji
  • Variational Depth Superresolution Using Example-Based Edge Representations [pdf] - David Ferstl, Matthias Ruther, Horst Bischof
  • Conformal and Low-Rank Sparse Representation for Image Restoration [pdf] - Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng
  • Learning Parametric Distributions for Image Super-Resolution: Where Patch Matching Meets Sparse Coding [pdf] - Yongbo Li, Weisheng Dong, Guangming Shi, Xuemei Xie
  • Convolutional Sparse Coding for Image Super-Resolution [pdf] - Shuhang Gu, Wangmeng Zuo, Qi Xie, Deyu Meng, Xiangchu Feng, Lei Zhang
Joint Object and Part Segmentation Using Deep Learned Potentials [pdf]
Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille

Abstract: Segmenting semantic objects from images and parsing them into their respective semantic parts are fundamental steps towards detailed object understanding in computer vision. In this paper, we propose a joint solution that tackles semantic object and part segmentation simultaneously, in which higher object-level context is provided to guide part segmentation, and more detailed part-level localization is utilized to refine object segmentation. Specifically, we first introduce the concept of semantic compositional parts (SCP) in which similar semantic parts are grouped and shared among different objects. A two-stream fully convolutional network (FCN) is then trained to provide the SCP and object potentials at each pixel. At the same time, a compact set of segments can also be obtained from the SCP predictions of the network. Given the potentials and the generated segments, in order to explore long-range context, we finally construct an efficient fully connected conditional random field (FCRF) to jointly predict the final object and part labels. Extensive evaluation on three different datasets shows that our approach can mutually enhance the performance of object and part segmentation, and outperforms the current state-of-the-art on both tasks.
Similar papers:
  • Learning Deconvolution Network for Semantic Segmentation [pdf] - Hyeonwoo Noh, Seunghoon Hong, Bohyung Han
  • Segment-Phrase Table for Semantic Segmentation, Visual Entailment and Paraphrasing [pdf] - Hamid Izadinia, Fereshteh Sadeghi, Santosh K. Divvala, Hannaneh Hajishirzi, Yejin Choi, Ali Farhadi
  • Semantic Segmentation With Object Clique Potential [pdf] - Xiaojuan Qi, Jianping Shi, Shu Liu, Renjie Liao, Jiaya Jia
  • Conditional Random Fields as Recurrent Neural Networks [pdf] - Shuai Zheng, Sadeep Jayasumana, Bernardino Romera-Paredes, Vibhav Vineet, Zhizhong Su, Dalong Du, Chang Huang, Philip H. S. Torr
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
Lost Shopping! Monocular Localization in Large Indoor Spaces [pdf]
Shenlong Wang, Sanja Fidler, Raquel Urtasun

Abstract: In this paper we propose a novel approach to localization in very large indoor spaces (i.e., shopping malls with over 200 stores) that takes a single image and a floor plan of the environment as input. We formulate the localization problem as inference in a Markov random field, which jointly reasons about text detection (localizing shop names in the image with precise bounding boxes), shop facade segmentation, as well as cameras rotation and translation within the entire shopping mall. The power of our approach is that it does not use any prior information about appearance and instead exploits text detections corresponding to shop names as a cue for localization. This makes our method applicable to a variety of domains and robust to store appearance variation across countries, seasons, and illumination conditions. We demonstrate our approach on our new dataset spanning two very large shopping malls, and show the power of holistic reasoning.
Similar papers:
  • Learning Informative Edge Maps for Indoor Scene Layout Prediction [pdf] - Arun Mallya, Svetlana Lazebnik
  • Localize Me Anywhere, Anytime: A Multi-Task Point-Retrieval Approach [pdf] - Guoyu Lu, Yan Yan, Li Ren, Jingkuan Song, Nicu Sebe, Chandra Kambhamettu
  • Text Flow: A Unified Text Detection System in Natural Scene Images [pdf] - Shangxuan Tian, Yifeng Pan, Chang Huang, Shijian Lu, Kai Yu, Chew Lim Tan
  • You Are Here: Mimicking the Human Thinking Process in Reading Floor-Plans [pdf] - Hang Chu, Dong Ki Kim, Tsuhan Chen
  • FASText: Efficient Unconstrained Scene Text Detector [pdf] - Michal Busta, Lukas Neumann, Jiri Matas
MMSS: Multi-Modal Sharable and Specific Feature Learning for RGB-D Object Recognition [pdf]
Anran Wang, Jianfei Cai, Jiwen Lu, Tat-Jen Cham

Abstract: Most of the feature-learning methods for RGB-D object recognition either learn features from color and depth modalities separately, or simply treat RGB-D as undifferentiated four-channel data, which cannot adequately exploit the relationship between different modalities. Motivated by the intuition that different modalities should contain not only some modal-specific patterns but also some shared common patterns, we propose a multi-modal feature learning framework for RGB-D object recognition. We first construct deep CNN layers for color and depth separately, and then connect them with our carefully designed multimodal layers, which fuse color and depth information by enforcing a common part to be shared by features of different modalities. In this way, we obtain features reflecting shared properties as well as modal-specific properties in different modalities. The information of the multi-modal learning frameworks is back-propagated to the early CNN layers. Experimental results show that our proposed multimodal feature learning method outperforms state-of-the-art approaches on two widely used RGB-D object benchmark datasets.
Similar papers:
  • Peeking Template Matching for Depth Extension [pdf] - Simon Korman, Eyal Ofek, Shai Avidan
  • High Quality Structure From Small Motion for Rolling Shutter Cameras [pdf] - Sunghoon Im, Hyowon Ha, Gyeongmin Choe, Hae-Gon Jeon, Kyungdon Joo, In So Kweon
  • Polarized 3D: High-Quality Depth Sensing With Polarization Cues [pdf] - Achuta Kadambi, Vage Taamazyan, Boxin Shi, Ramesh Raskar
  • Predicting Depth, Surface Normals and Semantic Labels With a Common Multi-Scale Convolutional Architecture [pdf] - David Eigen, Rob Fergus
  • Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf] - Naejin Kong, Michael J. Black
Mode-Seeking on Hypergraphs for Robust Geometric Model Fitting [pdf]
Hanzi Wang, Guobao Xiao, Yan Yan, David Suter

Abstract: In this paper, we propose a novel geometric model tting method, called Mode-Seeking on Hypergraphs (MSH), to deal with multi-structure data even in the presence of severe outliers. The proposed method formulates geometric model tting as a mode seeking problem on a hypergraph in which vertices represent model hypotheses and hyperedges denote data points. MSH intuitively detects model instances by a simple and effective mode seeking algorithm. In addition to the mode seeking algorithm, MSH includes a similarity measure between vertices on the hypergraph and a weight-aware sampling technique. The proposed method not only alleviates sensitivity to the data distribution, but also is scalable to large scale problems. Experimental results further demonstrate that the proposed method has signicant superiority over the state-of-the-art tting methods on both synthetic data and real images. (a) *equal contribution Weighting Score 0.05 0.04 0.03 0.02 0.01 0 1 1 0 p2 The minimum Tdistance (c)
Similar papers:
  • Globally Optimal 2D-3D Registration From Points or Lines Without Correspondences [pdf] - Mark Brown, David Windridge, Jean-Yves Guillemaut
  • Global, Dense Multiscale Reconstruction for a Billion Points [pdf] - Benjamin Ummenhofer, Thomas Brox
  • A Randomized Ensemble Approach to Industrial CT Segmentation [pdf] - Hyojin Kim, Jayaraman Jayaraman J. Thiagarajan, Peer-Timo Bremer
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • PatchMatch-Based Automatic Lattice Detection for Near-Regular Textures [pdf] - Siying Liu, Tian-Tsong Ng, Kalyan Sunkavalli, Minh N. Do, Eli Shechtman, Nathan Carr
Multiple Granularity Descriptors for Fine-Grained Categorization [pdf]
Dequan Wang, Zhiqiang Shen, Jie Shao, Wei Zhang, Xiangyang Xue, Zheng Zhang

Abstract: Fine-grained categorization, which aims to distinguish subordinate-level categories such as bird species or dog breeds, is an extremely challenging task. This is due to two main issues: how to localize discriminative regions for recognition and how to learn sophisticated features for representation. Neither of them is easy to handle if there is insufficient labeled data. We leverage the fact that a subordinate-level object already has other labels in its ontology tree. These free labels can be used to train a series of CNN-based classifiers, each specialized at one grain level. The internal representations of these networks have different region of interests, allowing the construction of multi-grained descriptors that encode informative and discriminative features covering all the grain levels. Our multiple granularity framework can be learned with the weakest supervision, requiring only image-level label and avoiding the use of labor-intensive bounding box or part annotations. Experimental results on three challenging fine-grained image datasets demonstrate that our approach outperforms state-of-the-art algorithms, including those requiring strong labels.
Similar papers:
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • Augmenting Strong Supervision Using Web Data for Fine-Grained Categorization [pdf] - Zhe Xu, Shaoli Huang, Ya Zhang, Dacheng Tao
  • Neural Activation Constellations: Unsupervised Part Model Discovery With Convolutional Networks [pdf] - Marcel Simon, Erik Rodner
  • Object Detection via a Multi-Region and Semantic Segmentation-Aware CNN Model [pdf] - Spyros Gidaris, Nikos Komodakis
  • Harvesting Discriminative Meta Objects With Deep CNN Features for Scene Classification [pdf] - Ruobing Wu, Baoyuan Wang, Wenping Wang, Yizhou Yu
Occlusion-Aware Depth Estimation Using Light-Field Cameras [pdf]
Ting-Chun Wang, Alexei A. Efros, Ravi Ramamoorthi

Abstract: Consumer-level and high-end light-field cameras are now widely available. Recent work has demonstrated practical methods for passive depth estimation from light-field images. However, most previous approaches do not explicitly model occlusions, and therefore cannot capture sharp transitions around object boundaries. A common assumption is that a pixel exhibits photo-consistency when focused to its correct depth, i.e., all viewpoints converge to a single (Lambertian) point in the scene. This assumption does not hold in the presence of occlusions, making most current approaches unreliable precisely where accurate depth information is most important at depth discontinuities. In this paper, we develop a depth estimation algorithm that treats occlusion explicitly; the method also enables identification of occlusion edges, which may be useful in other applications. We show that, although pixels at occlusions do not preserve photo-consistency in general, they are still consistent in approximately half the viewpoints. Moreover, the line separating the two view regions (correct depth vs. occluder) has the same orientation as the occlusion edge has in the spatial domain. By treating these two regions separately, depth estimation can be improved. Occlusion predictions can also be computed and used for regularization. Experimental results show that our method outperforms current state-of-the-art light-field depth estimation algorithms, especially near occlusion boundaries. Input light
Similar papers:
  • Beyond Tree Structure Models: A New Occlusion Aware Graphical Model for Human Pose Estimation [pdf] - Lianrui Fu, Junge Zhang, Kaiqi Huang
  • High Quality Structure From Small Motion for Rolling Shutter Cameras [pdf] - Sunghoon Im, Hyowon Ha, Gyeongmin Choe, Hae-Gon Jeon, Kyungdon Joo, In So Kweon
  • Depth Recovery From Light Field Using Focal Stack Symmetry [pdf] - Haiting Lin, Can Chen, Sing Bing Kang, Jingyi Yu
  • Large Displacement 3D Scene Flow With Occlusion Reasoning [pdf] - Andrei Zanfir, Cristian Sminchisescu
  • TransCut: Transparent Object Segmentation From a Light-Field Image [pdf] - Yichao Xu, Hajime Nagahara, Atsushi Shimada, Rin-ichiro Taniguchi
Photometric Stereo With Small Angular Variations [pdf]
Jian Wang, Yasuyuki Matsushita, Boxin Shi, Aswin C. Sankaranarayanan

Abstract: Most existing successful photometric stereo setups require large angular variations in illumination directions, which results in acquisition rigs that have large spatial extent. For many applications, especially involving mobile devices, it is important that the device be spatially compact. This naturally implies smaller angular variations in the illumination directions. This paper studies the effect of small angular variations in illumination directions to photometric stereo. We explore both theoretical justification and practical issues in the design of a compact and portable photometric stereo device on which a camera is surrounded by a ring of point light sources. We first derive the relationship between the estimation error of surface normal and the baseline of the point light sources. Armed with this theoretical insight, we develop a small baseline photometric stereo prototype to experimentally examine the theory and its practicality.
Similar papers:
  • Photogeometric Scene Flow for High-Detail Dynamic 3D Reconstruction [pdf] - Paulo F. U. Gotardo, Tomas Simon, Yaser Sheikh, Iain Matthews
  • On Linear Structure From Motion for Light Field Cameras [pdf] - Ole Johannsen, Antonin Sulc, Bastian Goldluecke
  • Nighttime Haze Removal With Glow and Multiple Light Colors [pdf] - Yu Li, Robby T. Tan, Michael S. Brown
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Photometric Stereo in a Scattering Medium [pdf] - Zak Murez, Tali Treibitz, Ravi Ramamoorthi, David Kriegman
Relaxed Multiple-Instance SVM With Application to Object Discovery [pdf]
Xinggang Wang, Zhuotun Zhu, Cong Yao, Xiang Bai

Abstract: Multiple-instance learning (MIL) has served as an important tool for a wide range of vision applications, for instance, image classification, object detection, and visual tracking. In this paper, we propose a novel method to solve the classical MIL problem, named relaxed multipleinstance SVM (RMI-SVM). We treat the positiveness of instance as a continuous variable, use Noisy-OR model to enforce the MIL constraints, and jointly optimize the bag label and instance label in a unified framework. The optimization problem can be efficiently solved using stochastic gradient decent. The extensive experiments demonstrate that RMI-SVM consistently achieves superior performance on various benchmarks for MIL. Moreover, we simply applied RMI-SVM to a challenging vision task, common object discovery. The state-of-the-art results of object discovery on Pascal VOC datasets further confirm the advantages of the proposed method.
Similar papers:
  • AttentionNet: Aggregating Weak Directions for Accurate Object Detection [pdf] - Donggeun Yoo, Sunggyun Park, Joon-Young Lee, Anthony S. Paek, In So Kweon
  • Towards Computational Baby Learning: A Weakly-Supervised Approach for Object Detection [pdf] - Xiaodan Liang, Si Liu, Yunchao Wei, Luoqi Liu, Liang Lin, Shuicheng Yan
  • Monocular Object Instance Segmentation and Depth Ordering With CNNs [pdf] - Ziyu Zhang, Alexander G. Schwing, Sanja Fidler, Raquel Urtasun
  • A Self-Paced Multiple-Instance Learning Framework for Co-Saliency Detection [pdf] - Dingwen Zhang, Deyu Meng, Chao Li, Lu Jiang, Qian Zhao, Junwei Han
  • Simpler Non-Parametric Methods Provide as Good or Better Results to Multiple-Instance Learning [pdf] - Ragav Venkatesan, Parag Chandakkar, Baoxin Li
TRIC-track: Tracking by Regression With Incrementally Learned Cascades [pdf]
Xiaomeng Wang, Michel Valstar, Brais Martinez, Muhammad Haris Khan, Tony Pridmore

Abstract: This paper proposes a novel approach to part-based tracking by replacing local matching of an appearance model by direct prediction of the displacement between local image patches and part locations. We propose to use cascaded regression with incremental learning to track generic objects without any prior knowledge of an objects structure or appearance. We exploit the spatial constraints between parts by implicitly learning the shape and deformation parameters of the object in an online fashion. We integrate a multiple temporal scale motion model to initialise our cascaded regression search close to the target and to allow it to cope with occlusions. Experimental results show that our tracker ranks first on the CVPR 2013 Benchmark. (a) (b) (c) (d) Figure 1: (a) Training a direct displacement regressor with four examples. (b) Testing a regressor. Four test patches sampled around the initial location (blue dot) provide predictions (purple dots). (c) Evidence aggregation map. (d) Location-based initialisation and implicit shape model.
Similar papers:
  • SOWP: Spatially Ordered and Weighted Patch Descriptor for Visual Tracking [pdf] - Han-Ul Kim, Dae-Youn Lee, Jae-Young Sim, Chang-Su Kim
  • Learning to Track: Online Multi-Object Tracking by Decision Making [pdf] - Yu Xiang, Alexandre Alahi, Silvio Savarese
  • Visual Tracking With Fully Convolutional Networks [pdf] - Lijun Wang, Wanli Ouyang, Xiaogang Wang, Huchuan Lu
  • Discriminative Low-Rank Tracking [pdf] - Yao Sui, Yafei Tang, Li Zhang
  • Understanding and Diagnosing Visual Tracking Systems [pdf] - Naiyan Wang, Jianping Shi, Dit-Yan Yeung, Jiaya Jia
Understanding and Diagnosing Visual Tracking Systems [pdf]
Naiyan Wang, Jianping Shi, Dit-Yan Yeung, Jiaya Jia

Abstract: Several benchmark datasets for visual tracking research have been created in recent years. Despite their usefulness, whether they are sufficient for understanding and diagnosing the strengths and weaknesses of different trackers remains questionable. To address this issue, we propose a framework by breaking a tracker down into five constituent parts, namely, motion model, feature extractor, observation model, model updater, and ensemble post-processor. We then conduct ablative experiments on each component to study how it affects the overall result. Surprisingly, our findings are discrepant with some common beliefs in the visual tracking research community. We find that the feature extractor plays the most important role in a tracker. On the other hand, although the observation model is the focus of many studies, we find that it often brings no significant improvement. Moreover, the motion model and model updater contain many details that could affect the result. Also, the ensemble post-processor can improve the result substantially when the constituent trackers have high diversity. Based on our findings, we put together some very elementary building blocks to give a basic tracker which is competitive in performance to the state-of-the-art trackers. We believe our framework can provide a solid baseline when conducting controlled experiments for visual tracking research.
Similar papers:
  • TRIC-track: Tracking by Regression With Incrementally Learned Cascades [pdf] - Xiaomeng Wang, Michel Valstar, Brais Martinez, Muhammad Haris Khan, Tony Pridmore
  • Hierarchical Convolutional Features for Visual Tracking [pdf] - Chao Ma, Jia-Bin Huang, Xiaokang Yang, Ming-Hsuan Yang
  • Multi-Kernel Correlation Filter for Visual Tracking [pdf] - Ming Tang, Jiayi Feng
  • Online Object Tracking With Proposal Selection [pdf] - Yang Hua, Karteek Alahari, Cordelia Schmid
  • Learning Spatially Regularized Correlation Filters for Visual Tracking [pdf] - Martin Danelljan, Gustav Hager, Fahad Shahbaz Khan, Michael Felsberg
Unsupervised Learning of Visual Representations Using Videos [pdf]
Xiaolong Wang, Abhinav Gupta

Abstract: Is strong supervision necessary for learning a good visual representation? Do we really need millions of semantically-labeled images to train a Convolutional Neural Network (CNN)? In this paper, we present a simple yet surprisingly powerful approach for unsupervised learning of CNN. Specifically, we use hundreds of thousands of unlabeled videos from the web to learn visual representations. Our key idea is that visual tracking provides the supervision. That is, two patches connected by a track should have similar visual representation in deep feature space since they probably belong to the same object or object part. We design a Siamese-triplet network with a ranking loss function to train this CNN representation. Without using a single image from ImageNet, just using 100K unlabeled videos and the VOC 2012 dataset, we train an ensemble of unsupervised networks that achieves 52% mAP (no bounding box regression). This performance comes tantalizingly close to its ImageNet-supervised counterpart, an ensemble which achieves a mAP of 54.4%. We also show that our unsupervised network can perform competitively in other tasks such as surface-normal estimation.
Similar papers:
  • Harvesting Discriminative Meta Objects With Deep CNN Features for Scene Classification [pdf] - Ruobing Wu, Baoyuan Wang, Wenping Wang, Yizhou Yu
  • Deep Multi-Patch Aggregation Network for Image Style, Aesthetics, and Quality Estimation [pdf] - Xin Lu, Zhe Lin, Xiaohui Shen, Radomir Mech, James Z. Wang
  • Learning Image Representations Tied to Ego-Motion [pdf] - Dinesh Jayaraman, Kristen Grauman
  • Local Convolutional Features With Unsupervised Training for Image Retrieval [pdf] - Mattis Paulin, Matthijs Douze, Zaid Harchaoui, Julien Mairal, Florent Perronin, Cordelia Schmid
  • Unsupervised Visual Representation Learning by Context Prediction [pdf] - Carl Doersch, Abhinav Gupta, Alexei A. Efros
Visual Tracking With Fully Convolutional Networks [pdf]
Lijun Wang, Wanli Ouyang, Xiaogang Wang, Huchuan Lu

Abstract: We propose a new approach for general object tracking with fully convolutional neural network. Instead of treating convolutional neural network (CNN) as a black-box feature extractor, we conduct in-depth study on the properties of CNN features offline pre-trained on massive image data and classification task on ImageNet. The discoveries motivate the design of our tracking system. It is found that convolutional layers in different levels characterize the target from different perspectives. A top layer encodes more semantic features and serves as a category detector, while a lower layer carries more discriminative information and can better separate the target from distracters with similar appearance. Both layers are jointly used with a switch mechanism during tracking. It is also found that for a tracking target, only a subset of neurons are relevant. A feature map selection method is developed to remove noisy and irrelevant feature maps, which can reduce computation redundancy and improve tracking accuracy. Extensive evaluation on the widely used tracking benchmark [36] shows that the proposed tacker outperforms the state-of-the-art significantly.
Similar papers:
  • Tracking-by-Segmentation With Online Gradient Boosting Decision Tree [pdf] - Jeany Son, Ilchae Jung, Kayoung Park, Bohyung Han
  • SOWP: Spatially Ordered and Weighted Patch Descriptor for Visual Tracking [pdf] - Han-Ul Kim, Dae-Youn Lee, Jae-Young Sim, Chang-Su Kim
  • Learning to Track: Online Multi-Object Tracking by Decision Making [pdf] - Yu Xiang, Alexandre Alahi, Silvio Savarese
  • Hierarchical Convolutional Features for Visual Tracking [pdf] - Chao Ma, Jia-Bin Huang, Xiaokang Yang, Ming-Hsuan Yang
  • Discriminative Low-Rank Tracking [pdf] - Yao Sui, Yafei Tang, Li Zhang
Learning to Track for Spatio-Temporal Action Localization [pdf]
Philippe Weinzaepfel, Zaid Harchaoui, Cordelia Schmid

Abstract: We propose an effective approach for spatio-temporal action localization in realistic videos. The approach first detects proposals at the frame-level and scores them with a combination of static and motion CNN features. It then tracks high-scoring proposals throughout the video using a tracking-by-detection approach. Our tracker relies simultaneously on instance-level and class-level detectors. The tracks are scored using a spatio-temporal motion histogram, a descriptor at the track level, in combination with the CNN features. Finally, we perform temporal localization of the action using a sliding-window approach at the track level. We present experimental results for spatiotemporal localization on the UCF-Sports, J-HMDB and UCF-101 action localization datasets, where our approach outperforms the state of the art with a margin of 15%, 7% and 12% respectively in mAP.
Similar papers:
  • Action Recognition by Hierarchical Mid-Level Action Elements [pdf] - Tian Lan, Yuke Zhu, Amir Roshan Zamir, Silvio Savarese
  • Actionness-Assisted Recognition of Actions [pdf] - Ye Luo, Loong-Fah Cheong, An Tran
  • Action Detection by Implicit Intentional Motion Clustering [pdf] - Wei Chen, Jason J. Corso
  • Action Localization in Videos Through Context Walk [pdf] - Khurram Soomro, Haroon Idrees, Mubarak Shah
  • Objects2action: Classifying and Localizing Actions Without Any Video Example [pdf] - Mihir Jain, Jan C. van Gemert, Thomas Mensink, Cees G. M. Snoek
Learning Concept Embeddings With Combined Human-Machine Expertise [pdf]
Michael Wilber, Iljung S. Kwak, David Kriegman, Serge Belongie

Abstract: This paper presents our work on SNaCK, a lowdimensional concept embedding algorithm that combines human expertise with automatic machine similarity kernels. Both parts are complimentary: human insight can capture relationships that are not apparent from the objects visual similarity and the machine can help relieve the human from having to exhaustively specify many constraints. We show that our SNaCK embeddings are useful in several tasks: distinguishing prime and nonprime numbers on MNIST, discovering labeling mistakes in the Caltech UCSD Birds (CUB) dataset with the help of deep-learned features, creating training datasets for bird classifiers, capturing subjective human taste on a new dataset of 10,000 foods, and qualitatively exploring an unstructured set of pictographic characters. Comparisons with the state-of-the-art in these tasks show that SNaCK produces better concept embeddings that require less human supervision than the leading methods. 1 Introduction Supervised learning tasks form the backbone of many stateof-the-art computer vision applications. They help researchers classify, localize, and characterize actions and objects. However, if the researchers goal is instead to interactively explore the latent structure of a dataset, discover novel categories, or find labeling mistakes, it is unclear what kind of supervision to use. Sometimes the data does not fall into well-defined taxonomic categories, or perhaps it is simply too expensive to collect labels fo
Similar papers:
  • BodyPrint: Pose Invariant 3D Shape Matching of Human Bodies [pdf] - Jiangping Wang, Kai Ma, Vivek Kumar Singh, Thomas Huang, Terrence Chen
  • Piecewise Flat Embedding for Image Segmentation [pdf] - Yizhou Yu, Chaowei Fang, Zicheng Liao
  • Semi-Supervised Zero-Shot Classification With Label Representation Learning [pdf] - Xin Li, Yuhong Guo, Dale Schuurmans
  • Automatic Concept Discovery From Parallel Text and Visual Corpora [pdf] - Chen Sun, Chuang Gan, Ram Nevatia
  • Im2Calories: Towards an Automated Mobile Vision Food Diary [pdf] - Austin Meyers, Nick Johnston, Vivek Rathod, Anoop Korattikara, Alex Gorban, Nathan Silberman, Sergio Guadarrama, George Papandreou, Jonathan Huang, Kevin P. Murphy
Depth Map Estimation and Colorization of Anaglyph Images Using Local Color Prior and Reverse Intensity Distribution [pdf]
Williem, Ramesh Raskar, In Kyu Park

Abstract: In this paper, we present a joint iterative anaglyph stereo matching and colorization framework for obtaining a set of disparity maps and colorized images. Conventional stereo matching algorithms fail when addressing anaglyph images that do not have similar intensities on their two respective view images. To resolve this problem, we propose two novel data costs using local color prior and reverse intensity distribution factor for obtaining accurate depth maps. To colorize an anaglyph image, each pixel in one view is warped to another view using the obtained disparity values of nonoccluded regions. A colorization algorithm using optimization is then employed with additional constraint to colorize the remaining occluded regions. Experimental results confirm that the proposed unified framework is robust and produces accurate depth maps and colorized stereo images.
Similar papers:
  • The HCI Stereo Metrics: Geometry-Aware Performance Analysis of Stereo Algorithms [pdf] - Katrin Honauer, Lena Maier-Hein, Daniel Kondermann
  • StereoSnakes: Contour Based Consistent Object Extraction For Stereo Images [pdf] - Ran Ju, Tongwei Ren, Gangshan Wu
  • Blur-Aware Disparity Estimation From Defocus Stereo Images [pdf] - Ching-Hui Chen, Hui Zhou, Timo Ahonen
  • MAP Disparity Estimation Using Hidden Markov Trees [pdf] - Eric T. Psota, Jedrzej Kowalczuk, Mateusz Mittek, Lance C. Perez
  • A Deep Visual Correspondence Embedding Model for Stereo Matching Costs [pdf] - Zhuoyuan Chen, Xun Sun, Liang Wang, Yinan Yu, Chang Huang
One Shot Learning via Compositions of Meaningful Patches [pdf]
Alex Wong, Alan L. Yuille

Abstract: Training Sample Perhaps one of the more impressive feats of human intelligence is the ability to learn a concept from few examples or even just one. At a young age, children easily learn their first language without complete exposure to the entire language and can even make inferences on novel concepts from their limited knowledge [8]. In fact, they can acquire a new word based on a single encounter [4]. However, if we survey state of the art learning methods, the results presented are the product of training from thousands of examples, where even a simple method such as logistic regression can perform very well [17]. Such performance becomes difficult to attain with only a single example. We believe that the basis for one-shot learning stems from the sharing of similar structures amongst objects of the same class. A bicycle can be parsed into a set of handles connected to the frame with two wheels and a seat. We can easily recognize a similar visual concept (eg. tricycle or motor bike) when it can be decomposed to a similar set of parts that embodies the structure of the objects. These parts
Similar papers:
  • Projection Onto the Manifold of Elongated Structures for Accurate Extraction [pdf] - Amos Sironi, Vincent Lepetit, Pascal Fua
  • External Patch Prior Guided Internal Clustering for Image Denoising [pdf] - Fei Chen, Lei Zhang, Huimin Yu
  • Learning Parametric Distributions for Image Super-Resolution: Where Patch Matching Meets Sparse Coding [pdf] - Yongbo Li, Weisheng Dong, Guangming Shi, Xuemei Xie
  • Unsupervised Visual Representation Learning by Context Prediction [pdf] - Carl Doersch, Abhinav Gupta, Alexei A. Efros
  • Harvesting Discriminative Meta Objects With Deep CNN Features for Scene Classification [pdf] - Ruobing Wu, Baoyuan Wang, Wenping Wang, Yizhou Yu
Rendering of Eyes for Eye-Shape Registration and Gaze Estimation [pdf]
Erroll Wood, Tadas Baltrusaitis, Xucong Zhang, Yusuke Sugano, Peter Robinson, Andreas Bulling

Abstract: Images of the eye are key in several computer vision problems, such as shape registration and gaze estimation. Recent large-scale supervised methods for these problems require time-consuming data collection and manual annotation, which can be unreliable. We propose synthesizing perfectly labelled photo-realistic training data in a fraction of the time. We used computer graphics techniques to build a collection of dynamic eye-region models from head scan geometry. These were randomly posed to synthesize close-up eye images for a wide range of head poses, gaze directions, and illumination conditions. We used our models controllability to verify the importance of realistic illumination and shape variations in eye-region training data. Finally, we demonstrate the benefits of our synthesized training data (SynthesEyes) by out-performing state-of-the-art methods for eye-shape registration as well as cross-dataset appearancebased gaze estimation in the wild.
Similar papers:
  • Single Image Pop-Up From Discriminatively Learned Parts [pdf] - Menglong Zhu, Xiaowei Zhou, Kostas Daniilidis
  • Robust Facial Landmark Detection Under Significant Head Poses and Occlusion [pdf] - Yue Wu, Qiang Ji
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
  • Fast and Accurate Head Pose Estimation via Random Projection Forests [pdf] - Donghoon Lee, Ming-Hsuan Yang, Songhwai Oh
  • Robust Model-Based 3D Head Pose Estimation [pdf] - Gregory P. Meyer, Shalini Gupta, Iuri Frosio, Dikpal Reddy, Jan Kautz
Wide-Area Image Geolocalization With Aerial Reference Imagery [pdf]
Scott Workman, Richard Souvenir, Nathan Jacobs

Abstract: We propose to use deep convolutional neural networks to address the problem of cross-view image geolocalization, in which the geolocation of a ground-level query image is estimated by matching to georeferenced aerial images. We use state-of-the-art feature representations for ground-level images and introduce a cross-view training approach for learning a joint semantic feature representation for aerial images. We also propose a network architecture that fuses features extracted from aerial images at multiple spatial scales. To support training these networks, we introduce a massive database that contains pairs of aerial and groundlevel images from across the United States. Our methods significantly out-perform the state of the art on two benchmark datasets. We also show, qualitatively, that the proposed feature representations are discriminative at both local and continental spatial scales.
Similar papers:
  • PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization [pdf] - Alex Kendall, Matthew Grimes, Roberto Cipolla
  • Webly Supervised Learning of Convolutional Networks [pdf] - Xinlei Chen, Abhinav Gupta
  • Localize Me Anywhere, Anytime: A Multi-Task Point-Retrieval Approach [pdf] - Guoyu Lu, Yan Yan, Li Ren, Jingkuan Song, Nicu Sebe, Chandra Kambhamettu
  • Semantically-Aware Aerial Reconstruction From Multi-Modal Data [pdf] - Randi Cabezas, Julian Straub, John W. Fisher III
  • Enhancing Road Maps by Parsing Aerial Images Around the World [pdf] - Gellert Mattyus, Shenlong Wang, Sanja Fidler, Raquel Urtasun
Adaptive Dither Voting for Robust Spatial Verification [pdf]
Xiaomeng Wu, Kunio Kashino

Abstract: Hough voting in a geometric transformation space allows us to realize spatial verification, but remains sensitive to feature detection errors because of the inflexible quantization of single feature correspondences. To handle this problem, we propose a new method, called adaptive dither voting, for robust spatial verification. For each correspondence, instead of hard-mapping it to a single transformation, the method augments its description by using multiple dithered transformations that are deterministically generated by the other correspondences. The method reduces the probability of losing correspondences during transformation quantization, and provides high robustness as regards mismatches by imposing three geometric constraints on the dithering process. We also propose exploiting the non-uniformity of a Hough histogram as the spatial similarity to handle multiple matching surfaces. Extensive experiments conducted on four datasets show the superiority of our method. The method outperforms its state-of-the-art counterparts in both accuracy and scalability, especially when it comes to the retrieval of small, rotated objects.
Similar papers:
  • Probabilistic Appearance Models for Segmentation and Classification [pdf] - Julia Kruger, Jan Ehrhardt, Heinz Handels
  • Person Re-Identification With Correspondence Structure Learning [pdf] - Yang Shen, Weiyao Lin, Junchi Yan, Mingliang Xu, Jianxin Wu, Jingdong Wang
  • Dense Semantic Correspondence Where Every Pixel is a Classifier [pdf] - Hilton Bristow, Jack Valmadre, Simon Lucey
  • Hyperpoints and Fine Vocabularies for Large-Scale Location Recognition [pdf] - Torsten Sattler, Michal Havlena, Filip Radenovic, Konrad Schindler, Marc Pollefeys
  • Camera Pose Voting for Large-Scale Image-Based Localization [pdf] - Bernhard Zeisl, Torsten Sattler, Marc Pollefeys
Harvesting Discriminative Meta Objects With Deep CNN Features for Scene Classification [pdf]
Ruobing Wu, Baoyuan Wang, Wenping Wang, Yizhou Yu

Abstract: Recent work on scene classification still makes use of generic CNN features in a rudimentary manner. In this paper, we present a novel pipeline built upon deep CNN features to harvest discriminative visual objects and parts for scene classification. We first use a region proposal technique to generate a set of high-quality patches potentially containing objects, and apply a pre-trained CNN to extract generic deep features from these patches. Then we perform both unsupervised and weakly supervised learning to screen these patches and discover discriminative ones representing category-specific objects and parts. We further apply discriminative clustering enhanced with local CNN finetuning to aggregate similar objects and parts into groups, called meta objects. A scene image representation is constructed by pooling the feature response maps of all the learned meta objects at multiple spatial scales. We have confirmed that the scene image representation obtained using this new pipeline is capable of delivering state-of-theart performance on two popular scene benchmark datasets, MIT Indoor 67 [22] and Sun397 [31].
Similar papers:
  • BubbLeNet: Foveated Imaging for Visual Discovery [pdf] - Kevin Matzen, Noah Snavely
  • Local Convolutional Features With Unsupervised Training for Image Retrieval [pdf] - Mattis Paulin, Matthijs Douze, Zaid Harchaoui, Julien Mairal, Florent Perronin, Cordelia Schmid
  • One Shot Learning via Compositions of Meaningful Patches [pdf] - Alex Wong, Alan L. Yuille
  • External Patch Prior Guided Internal Clustering for Image Denoising [pdf] - Fei Chen, Lei Zhang, Huimin Yu
  • Unsupervised Visual Representation Learning by Context Prediction [pdf] - Carl Doersch, Abhinav Gupta, Alexei A. Efros
ML-MG: Multi-Label Learning With Missing Labels Using a Mixed Graph [pdf]
Baoyuan Wu, Siwei Lyu, Bernard Ghanem

Abstract: This work focuses on the problem of multi-label learning with missing labels (MLML), which aims to label each test instance with multiple class labels given training instances that have an incomplete/partial set of these labels (i.e. some of their labels are missing). To handle missing labels, we propose a unied model of label dependencies by constructing a mixed graph, which jointly incorporates (i) instancelevel similarity and class co-occurrence as undirected edges and (ii) semantic label hierarchy as directed edges. Unlike most MLML methods, We formulate this learning problem transductively as a convex quadratic matrix optimization problem that encourages training label consistency and encodes both types of label dependencies (i.e. undirected and directed edges) using quadratic terms and hard linear constraints. The alternating direction method of multipliers (ADMM) can be used to exactly and efciently solve this problem. To evaluate our proposed method, we consider two popular applications (image and video annotation), where the label hierarchy can be derived from Wordnet. Experimental results show that our method achieves a signicant improvement over state-of-the-art methods in performance and robustness to missing labels.
Similar papers:
  • Semantically-Aware Aerial Reconstruction From Multi-Modal Data [pdf] - Randi Cabezas, Julian Straub, John W. Fisher III
  • Learning Semi-Supervised Representation Towards a Unified Optimization Framework for Semi-Supervised Learning [pdf] - Chun-Guang Li, Zhouchen Lin, Honggang Zhang, Jun Guo
  • Probabilistic Label Relation Graphs With Ising Models [pdf] - Nan Ding, Jia Deng, Kevin P. Murphy, Hartmut Neven
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • Semi-Supervised Zero-Shot Classification With Label Representation Learning [pdf] - Xin Li, Yuhong Guo, Dale Schuurmans
Robust Facial Landmark Detection Under Significant Head Poses and Occlusion [pdf]
Yue Wu, Qiang Ji

Abstract: There have been tremendous improvements for facial landmark detection on general in-the-wild images. However, it is still challenging to detect the facial landmarks on images with severe occlusion and images with large head poses (e.g. profile face). In fact, the existing algorithms usually can only handle one of them. In this work, we propose a unified robust cascade regression framework that can handle both images with severe occlusion and images with large head poses. Specifically, the method iteratively predicts the landmark occlusions and the landmark locations. For occlusion estimation, instead of directly predicting the binary occlusion vectors, we introduce a supervised regression method that gradually updates the landmark visibility probabilities in each iteration to achieve robustness. In addition, we explicitly add occlusion pattern as a constraint to improve the performance of occlusion prediction. For landmark detection, we combine the landmark visibility probabilities, the local appearances, and the local shapes to iteratively update their positions. The experimental results show that the proposed method is significantly better than state-of-the-art works on images with severe occlusion and images with large head poses. It is also comparable to other methods on general in-the-wild images.
Similar papers:
  • Joint Fine-Tuning in Deep Neural Networks for Facial Expression Recognition [pdf] - Heechul Jung, Sihaeng Lee, Junho Yim, Sunjeong Park, Junmo Kim
  • Scene-Domain Active Part Models for Object Representation [pdf] - Zhou Ren, Chaohui Wang, Alan L. Yuille
  • Regressive Tree Structured Model for Facial Landmark Localization [pdf] - Gee-Sern Hsu, Kai-Hsiang Chang, Shih-Chieh Huang
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • Single Image Pop-Up From Discriminatively Learned Parts [pdf] - Menglong Zhu, Xiaowei Zhou, Kostas Daniilidis
Video Restoration Against Yin-Yang Phasing [pdf]
Xiaolin Wu, Zhenhao Li, Xiaowei Deng

Abstract: A common video degradation problem, which is largely untreated in literature, is what we call Yin-Yang Phasing (YYP). YYP is characterized by involuntary, dramatic flipflop in the intensity and possibly chromaticity of an object as the video plays. Such temporal artifacts occur under ill illumination conditions and are triggered by object or/and camera motions, which mislead the settings of cameras auto-exposure and white point. In this paper, we investigate the problem and propose a video restoration technique to suppress YYP artifacts and retain temporal consistency of objects appearance via inter-frame, spatially-adaptive, optimal tone mapping. The video quality can be further improved by a novel image enhancer designed in Webers perception principle and by exploiting the second-order statistics of the scene. Experimental results are encouraging, pointing to an effective, practical solution for a common but surprisingly understudied problem.
Similar papers:
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
Learning Discriminative Reconstructions for Unsupervised Outlier Removal [pdf]
Yan Xia, Xudong Cao, Fang Wen, Gang Hua, Jian Sun

Abstract: We study the problem of automatically removing outliers from noisy data, with application for removing outlier images from an image collection. We address this problem by utilizing the reconstruction errors of an autoencoder. We observe that when data are reconstructed from lowdimensional representations, the inliers and the outliers can be well separated according to their reconstruction errors. Based on this basic observation, we gradually inject discriminative information in the learning process of autoencoder to make the inliers and the outliers more separable. Experiments on a variety of image datasets validate our approach.
Similar papers:
  • Robust and Optimal Sum-of-Squares-Based Point-to-Plane Registration of Image Sets and Structured Scenes [pdf] - Danda Pani Paudel, Adlane Habed, Cedric Demonceaux, Pascal Vasseur
  • Harvesting Discriminative Meta Objects With Deep CNN Features for Scene Classification [pdf] - Ruobing Wu, Baoyuan Wang, Wenping Wang, Yizhou Yu
  • Flow Fields: Dense Correspondence Fields for Highly Accurate Large Displacement Optical Flow Estimation [pdf] - Christian Bailer, Bertram Taetz, Didier Stricker
  • Domain Generalization for Object Recognition With Multi-Task Autoencoders [pdf] - Muhammad Ghifary, W. Bastiaan Kleijn, Mengjie Zhang, David Balduzzi
  • Guaranteed Outlier Removal for Rotation Search [pdf] - Alvaro Parra Bustos, Tat-Jun Chin
Learning to Track: Online Multi-Object Tracking by Decision Making [pdf]
Yu Xiang, Alexandre Alahi, Silvio Savarese

Abstract: Online Multi-Object Tracking (MOT) has wide applications in time-critical video analysis scenarios, such as robot navigation and autonomous driving. In trackingby-detection, a major challenge of online MOT is how to robustly associate noisy object detections on a new video frame with previously tracked objects. In this work, we formulate the online MOT problem as decision making in Markov Decision Processes (MDPs), where the lifetime of an object is modeled with a MDP. Learning a similarity function for data association is equivalent to learning a policy for the MDP, and the policy learning is approached in a reinforcement learning fashion which benefits from both advantages of offline-learning and online-learning for data association. Moreover, our framework can naturally handle the birth/death and appearance/disappearance of targets by treating them as state transitions in the MDP while leveraging existing online single object tracking methods. We conduct experiments on the MOT Benchmark [24] to verify the effectiveness of our method.
Similar papers:
  • Discriminative Low-Rank Tracking [pdf] - Yao Sui, Yafei Tang, Li Zhang
  • Tracking-by-Segmentation With Online Gradient Boosting Decision Tree [pdf] - Jeany Son, Ilchae Jung, Kayoung Park, Bohyung Han
  • Learning to Divide and Conquer for Online Multi-Target Tracking [pdf] - Francesco Solera, Simone Calderara, Rita Cucchiara
  • Joint Probabilistic Data Association Revisited [pdf] - Seyed Hamid Rezatofighi, Anton Milan, Zhen Zhang, Qinfeng Shi, Anthony Dick, Ian Reid
  • Near-Online Multi-Target Tracking With Aggregated Local Flow Descriptor [pdf] - Wongun Choi
Discovering the Spatial Extent of Relative Attributes [pdf]
Fanyi Xiao, Yong Jae Lee

Abstract: We present a weakly-supervised approach that discovers the spatial extent of relative attributes, given only pairs of ordered images. In contrast to traditional approaches that use global appearance features or rely on keypoint detectors, our goal is to automatically discover the image regions that are relevant to the attribute, even when the attributes appearance changes drastically across its attribute spectrum. To accomplish this, we rst develop a novel formulation that combines a detector with local smoothness to discover a set of coherent visual chains across the image collection. We then introduce an efcient way to generate additional chains anchored on the initial discovered ones. Finally, we automatically identify the most relevant visual chains, and create an ensemble image representation to model the attribute. Through extensive experiments, we demonstrate our methods promise relative to several baselines in modeling relative attributes.
Similar papers:
  • Multi-Task Learning With Low Rank Attribute Embedding for Person Re-Identification [pdf] - Chi Su, Fan Yang, Shiliang Zhang, Qi Tian, Larry S. Davis, Wen Gao
  • Attributed Grammars for Joint Estimation of Human Attributes, Part and Pose [pdf] - Seyoung Park, Song-Chun Zhu
  • Learning Deep Representation With Large-Scale Attributes [pdf] - Wanli Ouyang, Hongyang Li, Xingyu Zeng, Xiaogang Wang
  • Just Noticeable Differences in Visual Attributes [pdf] - Aron Yu, Kristen Grauman
  • A Unified Multiplicative Framework for Attribute Learning [pdf] - Kongming Liang, Hong Chang, Shiguang Shan, Xilin Chen
Holistically-Nested Edge Detection [pdf]
Saining Xie, Zhuowen Tu

Abstract: We develop a new edge detection algorithm that addresses two important issues in this long-standing vision problem: (1) holistic image training and prediction; and (2) multi-scale and multi-level feature learning. Our proposed method, holistically-nested edge detection (HED), performs image-to-image prediction by means of a deep learning model that leverages fully convolutional neural networks and deeply-supervised nets. HED automatically learns rich hierarchical representations (guided by deep supervision on side responses) that are important in order to resolve the challenging ambiguity in edge and object boundary detection. We significantly advance the state-of-the-art on the BSD500 dataset (ODS F-score of .782) and the NYU Depth dataset (ODS F-score of .746), and do so with an improved speed (0.4s per image) that is orders of magnitude faster than some recent CNN-based edge detection algorithms.
Similar papers:
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • Learning The Structure of Deep Convolutional Networks [pdf] - Jiashi Feng, Trevor Darrell
  • Learning Informative Edge Maps for Indoor Scene Layout Prediction [pdf] - Arun Mallya, Svetlana Lazebnik
  • Realtime Edge-Based Visual Odometry for a Monocular Camera [pdf] - Juan Jose Tarrio, Sol Pedre
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
RIDE: Reversal Invariant Descriptor Enhancement [pdf]
Lingxi Xie, Jingdong Wang, Weiyao Lin, Bo Zhang, Qi Tian

Abstract: In many fine-grained object recognition datasets, image orientation (left/right) might vary from sample to sample. Since handcrafted descriptors such as SIFT are not reversal invariant, the stability of image representation based on them is consequently limited. A popular solution is to augment the datasets by adding a left-right reversed copy for each original image. This strategy improves recognition accuracy to some extent, but also brings the price of almost doubled time and memory consumptions. In this paper, we present RIDE (Reversal Invariant Descriptor Enhancement) for fine-grained object recognition. RIDE is a generalized algorithm which cancels out the impact of image reversal by estimating the orientation of local descriptors, and guarantees to produce the identical representation for an image and its left-right reversed copy. Experimental results reveal the consistent accuracy gain of RIDE with various types of descriptors. We also provide insightful discussions on the working mechanism of RIDE and its generalization to other applications. SIFT SIFT+RIDE = . SIFT SIFT+RIDE = .
Similar papers:
  • Multi-View Convolutional Neural Networks for 3D Shape Recognition [pdf] - Hang Su, Subhransu Maji, Evangelos Kalogerakis, Erik Learned-Miller
  • Predicting Good Features for Image Geo-Localization Using Per-Bundle VLAD [pdf] - Hyo Jin Kim, Enrique Dunn, Jan-Michael Frahm
  • Aggregating Local Deep Features for Image Retrieval [pdf] - Artem Babenko, Victor Lempitsky
  • Local Convolutional Features With Unsupervised Training for Image Retrieval [pdf] - Mattis Paulin, Matthijs Douze, Zaid Harchaoui, Julien Mairal, Florent Perronin, Cordelia Schmid
  • Discriminative Learning of Deep Convolutional Feature Point Descriptors [pdf] - Edgar Simo-Serra, Eduard Trulls, Luis Ferraz, Iasonas Kokkinos, Pascal Fua, Francesc Moreno-Noguer
Task-Driven Feature Pooling for Image Classification [pdf]
Guo-Sen Xie, Xu-Yao Zhang, Xiangbo Shu, Shuicheng Yan, Cheng-Lin Liu

Abstract: 90 120 60 30 150 Feature pooling is an important strategy to achieve high performance in image classification. However, most pooling methods are unsupervised and heuristic. In this paper, we propose a novel task-driven pooling (TDP) model to directly learn the pooled representation from data in a discriminative manner. Different from the traditional methods (e.g., average and max pooling), TDP is an implicit pooling method which elegantly integrates the learning of representations into the given classification task. The optimization of TDP can equalize the similarities between the descriptors and the learned representation, and maximize the classification accuracy. TDP can be combined with the traditional BoW models (coding vectors) or the recent stateof-the-art CNN models (feature maps) to achieve a much better pooled representation. Furthermore, a self-training mechanism is used to generate the TDP representation for a new test image. A multi-task extension of TDP is also proposed to further improve the performance. Experiments on three databases (Flower-17, Indoor-67 and Caltech-101) well validate the effectiveness of our models.
Similar papers:
  • Flowing ConvNets for Human Pose Estimation in Videos [pdf] - Tomas Pfister, James Charles, Andrew Zisserman
  • Multi-Scale Recognition With DAG-CNNs [pdf] - Songfan Yang, Deva Ramanan
  • Aggregating Local Deep Features for Image Retrieval [pdf] - Artem Babenko, Victor Lempitsky
  • Fast R-CNN [pdf] - Ross Girshick
  • Unsupervised Learning of Spatiotemporally Coherent Metrics [pdf] - Ross Goroshin, Joan Bruna, Jonathan Tompson, David Eigen, Yann LeCun
Conditional Convolutional Neural Network for Modality-Aware Face Recognition [pdf]
Chao Xiong, Xiaowei Zhao, Danhang Tang, Karlekar Jayashree, Shuicheng Yan, Tae-Kyun Kim

Abstract: Faces in the wild are usually captured with various poses, illuminations and occlusions, and thus inherently multimodally distributed in many tasks. We propose a conditional Convolutional Neural Network, named as c-CNN, to handle multimodal face recognition. Different from traditional CNN that adopts fixed convolution kernels, samples in c-CNN are processed with dynamically activated sets of kernels. In particular, convolution kernels within each layer are only sparsely activated when a sample is passed through the network. For a given sample, the activations of convolution kernels in a certain layer are conditioned on its present intermediate representation and the activation status in the lower layers. The activated kernels across layers define the sample-specific adaptive routes that reveal the distribution of underlying modalities. Consequently, the proposed framework does not rely on any prior knowledge of modalities in contrast with most existing methods. To substantiate the generic framework, we introduce a special case of c-CNN via incorporating the conditional routing of the decision tree, which is evaluated with two problems of multimodality multi-view face identification and occluded face verification. Extensive experiments demonstrate consistent improvements over the counterparts unaware of modalities.
Similar papers:
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • Personalized Age Progression With Aging Dictionary [pdf] - Xiangbo Shu, Jinhui Tang, Hanjiang Lai, Luoqi Liu, Shuicheng Yan
  • Simultaneous Local Binary Feature Learning and Encoding for Face Recognition [pdf] - Jiwen Lu, Venice Erin Liong, Jie Zhou
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Deep Neural Decision Forests [pdf] - Peter Kontschieder, Madalina Fiterau, Antonio Criminisi, Samuel Rota Bulo
Storyline Representation of Egocentric Videos With an Applications to Story-Based Search [pdf]
Bo Xiong, Gunhee Kim, Leonid Sigal

Abstract: Egocentric videos are a valuable source of information as a daily log of our lives. However, large fraction of egocentric video content is typically irrelevant and boring to re-watch. It is an agonizing task, for example, to manually search for the moment when your daughter first met Mickey Mouse from hours-long egocentric videos taken at Disneyland. Although many summarization methods have been successfully proposed to create concise representations of videos, in practice, the value of the subshots to users may change according to their immediate preference/mood; thus summaries with fixed criteria may not fully satisfy users various search intents. To address this, we propose a storyline representation that expresses an egocentric video as a set of jointly inferred, through MRF inference, story elements comprising of actors, locations, supporting objects and events, depicted on a timeline. We construct such a storyline with very limited annotation data (a list of map locations and weak knowledge of what events may be possible at each location), by bootstrapping the process with data obtained through focused Web image and video searches. Our representation promotes story-based search with queries in the form of AND-OR graphs, which span any subset of story elements and their spatio-temporal composition. We show effectiveness of our approach on a set of unconstrained YouTube egocentric videos of visits to Disneyland.
Similar papers:
  • Selecting Relevant Web Trained Concepts for Automated Event Retrieval [pdf] - Bharat Singh, Xintong Han, Zhe Wu, Vlad I. Morariu, Larry S. Davis
  • Temporal Perception and Prediction in Ego-Centric Video [pdf] - Yipin Zhou, Tamara L. Berg
  • Lending A Hand: Detecting Hands and Recognizing Activities in Complex Egocentric Interactions [pdf] - Sven Bambach, Stefan Lee, David J. Crandall, Chen Yu
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
Activity Auto-Completion: Predicting Human Activities From Partial Videos [pdf]
Zhen Xu, Laiyun Qing, Jun Miao

Abstract: In this paper, we propose an activity auto-completion (AAC) model for human activity prediction by formulating activity prediction as a query auto-completion (QAC) problem in information retrieval. First, we extract discriminative patches in frames of videos. A video is represented based on these patches and divided into a collection of segments, each of which is regarded as a character typed in the search box. Then a partially observed video is considered as an activity prex, consisting of one or more characters. Finally, the missing observation of an activity is predicted as the activity candidates provided by the auto-completion model. The candidates are matched against the activity prex on-the-fly and ranked by a learning-to-rank algorithm. We validate our method on UT-Interaction Set #1 and Set #2 [19]. The experimental results show that the proposed activity auto-completion model achieves promising performance.
Similar papers:
  • Learning Ensembles of Potential Functions for Structured Prediction With Latent Variables [pdf] - Hossein Hajimirsadeghi, Greg Mori
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Temporal Perception and Prediction in Ego-Centric Video [pdf] - Yipin Zhou, Tamara L. Berg
  • Context Aware Active Learning of Activity Recognition Models [pdf] - Mahmudul Hasan, Amit K. Roy-Chowdhury
  • Unsupervised Semantic Parsing of Video Collections [pdf] - Ozan Sener, Amir R. Zamir, Silvio Savarese, Ashutosh Saxena
Augmenting Strong Supervision Using Web Data for Fine-Grained Categorization [pdf]
Zhe Xu, Shaoli Huang, Ya Zhang, Dacheng Tao

Abstract: We propose a new method for fine-grained object recognition that employs part-level annotations and deep convolutional neural networks (CNNs) in a unified framework. Although both schemes have been widely used to boost recognition performance, due to the difficulty in acquiring detailed part annotations, strongly supervised fine-grained datasets are usually too small to keep pace with the rapid evolution of CNN architectures. In this paper, we solve this problem by exploiting inexhaustible web data. The proposed method improves classification accuracy in two ways: more discriminative CNN feature representations are generated using a training set augmented by collecting a large number of part patches from weakly supervised web images; and more robust object classifiers are learned using a multi-instance learning algorithm jointly on the strong and weak datasets. Despite its simplicity, the proposed method delivers a remarkable performance improvement on the CUB200-2011 dataset compared to baseline part-based R-CNN methods, and achieves the highest accuracy on this dataset even in the absence of test image annotations.
Similar papers:
  • Webly Supervised Learning of Convolutional Networks [pdf] - Xinlei Chen, Abhinav Gupta
  • Harvesting Discriminative Meta Objects With Deep CNN Features for Scene Classification [pdf] - Ruobing Wu, Baoyuan Wang, Wenping Wang, Yizhou Yu
  • Neural Activation Constellations: Unsupervised Part Model Discovery With Convolutional Networks [pdf] - Marcel Simon, Erik Rodner
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • Weakly- and Semi-Supervised Learning of a Deep Convolutional Network for Semantic Image Segmentation [pdf] - George Papandreou, Liang-Chieh Chen, Kevin P. Murphy, Alan L. Yuille
Deformable 3D Fusion: From Partial Dynamic 3D Observations to Complete 4D Models [pdf]
Weipeng Xu, Mathieu Salzmann, Yongtian Wang, Yue Liu

Abstract: Capturing the 3D motion of dynamic, non-rigid objects has attracted significant attention in computer vision. Existing methods typically require either mostly complete 3D volumetric observations, or a shape template. In this paper, we introduce a template-less 4D reconstruction method that incrementally fuses highly-incomplete 3D observations of a deforming object, and generates a complete, temporallycoherent shape representation of the object. To this end, we design an online algorithm that alternatively registers new observations to the current model estimate and updates the model. We demonstrate the effectiveness of our approach at reconstructing non-rigidly moving objects from highlyincomplete measurements on both sequences of partial 3D point clouds and Kinect videos. Figure 1. Deformable 3D fusion.
Similar papers:
  • Discriminative Low-Rank Tracking [pdf] - Yao Sui, Yafei Tang, Li Zhang
  • Multi-View Subspace Clustering [pdf] - Hongchang Gao, Feiping Nie, Xuelong Li, Heng Huang
  • Direct, Dense, and Deformable: Template-Based Non-Rigid 3D Reconstruction From RGB Video [pdf] - Rui Yu, Chris Russell, Neill D. F. Campbell, Lourdes Agapito
  • Exploiting Object Similarity in 3D Reconstruction [pdf] - Chen Zhou, Fatma Guney, Yizhou Wang, Andreas Geiger
  • Shape Interaction Matrix Revisited and Robustified: Efficient Subspace Clustering With Corrupted and Incomplete Data [pdf] - Pan Ji, Mathieu Salzmann, Hongdong Li
Learning to Predict Saliency on Face Images [pdf]
Mai Xu, Yun Ren, Zulin Wang

Abstract: This paper proposes a novel method, which learns to detect saliency of face images. To be more specific, we obtain a database of eye tracking over extensive face images, via conducting an eye tracking experiment. With analysis on eye tracking database, we verify that the fixations tend to cluster around facial features, when viewing images with large faces. For modeling attention on faces and facial features, the proposed method learns the Gaussian mixture model (GMM) distribution from the fixations of eye tracking data as the top-down features for saliency detection of face images. Then, in our method, the top-down features (i.e., face and facial features) upon the the learnt GMM are linearly combined with the conventional bottom-up features (i.e., color, intensity, and orientation), for saliency detection. In the linear combination, we argue that the weights corresponding to top-down feature channels depend on the face size in images, and the relationship between the weights and face size is thus investigated via learning from the training eye tracking data. Finally, experimental results show that our learning-based method is able to advance state-of-the-art saliency prediction for face images. The corresponding database and code are available online: www.ee.buaa.edu.cn/xumfiles/saliency detection.html.
Similar papers:
  • Personalized Age Progression With Aging Dictionary [pdf] - Xiangbo Shu, Jinhui Tang, Hanjiang Lai, Luoqi Liu, Shuicheng Yan
  • A Data-Driven Metric for Comprehensive Evaluation of Saliency Models [pdf] - Jia Li, Changqun Xia, Yafei Song, Shu Fang, Xiaowu Chen
  • Cluster-Based Point Set Saliency [pdf] - Flora Ponjou Tasse, Jiri Kosinka, Neil Dodgson
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf] - Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao
Patch Group Based Nonlocal Self-Similarity Prior Learning for Image Denoising [pdf]
Jun Xu, Lei Zhang, Wangmeng Zuo, David Zhang, Xiangchu Feng

Abstract: Patch based image modeling has achieved a great success in low level vision such as image denoising. In particular, the use of image nonlocal self-similarity (NSS) prior, which refers to the fact that a local patch often has many nonlocal similar patches to it across the image, has significantly enhanced the denoising performance. However, in most existing methods only the NSS of input degraded image is exploited, while how to utilize the NSS of clean natural images is still an open problem. In this paper, we propose a patch group (PG) based NSS prior learning scheme to learn explicit NSS models from natural images for high performance denoising. PGs are extracted from training images by putting nonlocal similar patches into groups, and a PG based Gaussian Mixture Model (PG-GMM) learning algorithm is developed to learn the NSS prior. We demonstrate that, owe to the learned PG-GMM, a simple weighted sparse coding model, which has a closed-form solution, can be used to perform image denoising effectively, resulting in high PSNR measure, fast speed, and particularly the best visual quality among all competing methods.
Similar papers:
  • Learning Nonlinear Spectral Filters for Color Image Reconstruction [pdf] - Michael Moeller, Julia Diebold, Guy Gilboa, Daniel Cremers
  • Learning Parametric Distributions for Image Super-Resolution: Where Patch Matching Meets Sparse Coding [pdf] - Yongbo Li, Weisheng Dong, Guangming Shi, Xuemei Xie
  • Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf] - Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • Low-Rank Tensor Approximation With Laplacian Scale Mixture Modeling for Multiframe Image Denoising [pdf] - Weisheng Dong, Guangyu Li, Guangming Shi, Xin Li, Yi Ma
  • External Patch Prior Guided Internal Clustering for Image Denoising [pdf] - Fei Chen, Lei Zhang, Huimin Yu
TransCut: Transparent Object Segmentation From a Light-Field Image [pdf]
Yichao Xu, Hajime Nagahara, Atsushi Shimada, Rin-ichiro Taniguchi

Abstract: The segmentation of transparent objects can be very useful in computer vision applications. However, because they borrow texture from their background and have a similar appearance to their surroundings, transparent objects are not handled well by regular image segmentation methods. We propose a method that overcomes these problems using the consistency and distortion properties of a light-field image. Graph-cut optimization is applied for the pixel labeling problem. The light-field linearity is used to estimate the likelihood of a pixel belonging to the transparent object or Lambertian background, and the occlusion detector is used to find the occlusion boundary. We acquire a light field dataset for the transparent object, and use this dataset to evaluate our method. The results demonstrate that the proposed method successfully segments transparent objects from the background. Figure 1. Input and output of our system. The left side shows the captured light-field image, and the right-hand side is a magnification of the central viewpoint. The output after segmentation of the transparent object is shown on the bottom-right. bowls, bottles, and jars. One way of detecting these transparent objects is to segment them from captured images of the scene. The appearance of a transparent object is highly dependent on the background, from which its texture and colors are largely borrowed. Thus, it is extremely challenging to separate the transparent object from the background.
Similar papers:
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
  • Robust Facial Landmark Detection Under Significant Head Poses and Occlusion [pdf] - Yue Wu, Qiang Ji
  • Large Displacement 3D Scene Flow With Occlusion Reasoning [pdf] - Andrei Zanfir, Cristian Sminchisescu
  • Beyond Tree Structure Models: A New Occlusion Aware Graphical Model for Human Pose Estimation [pdf] - Lianrui Fu, Junge Zhang, Kaiqi Huang
  • Occlusion-Aware Depth Estimation Using Light-Field Cameras [pdf] - Ting-Chun Wang, Alexei A. Efros, Ravi Ramamoorthi
Unsupervised Trajectory Clustering via Adaptive Multi-Kernel-Based Shrinkage [pdf]
Hongteng Xu, Yang Zhou, Weiyao Lin, Hongyuan Zha

Abstract: This paper proposes a shrinkage-based framework for unsupervised trajectory clustering. Facing to the challenges of trajectory clustering, e.g., large variations within a cluster and ambiguities across clusters, we first introduce an adaptive multi-kernel-based estimation process to estimate the shrunk positions and speeds of trajectories points. This kernel-based estimation effectively leverages both multiple structural information within a trajectory and the local motion patterns across multiple trajectories, such that the discrimination of the shrunk point can be properly increased. We further introduce a speed-regularized optimization process, which utilizes the estimated speeds to regularize the optimal shrunk points, so as to guarantee the smoothness and the discriminative pattern of the final shrunk trajectory. Using our approach, the variations among similar trajectories can be reduced while the boundaries between different clusters are enlarged. Experimental results demonstrate that our approach is superior to the state-of-art approaches on both clustering accuracy and robustness. Besides, additional experiments reveal the effectiveness of our approach when applied to trajectory analysis applications, e.g., anomaly detection and route analysis.
Similar papers:
  • Dense Continuous-Time Tracking and Mapping With Rolling Shutter RGB-D Cameras [pdf] - Christian Kerl, Jorg Stuckler, Daniel Cremers
  • Learning Shape, Motion and Elastic Models in Force Space [pdf] - Antonio Agudo, Francesc Moreno-Noguer
  • Panoptic Studio: A Massively Multiview System for Social Motion Capture [pdf] - Hanbyul Joo, Hao Liu, Lei Tan, Lin Gui, Bart Nabbe, Iain Matthews, Takeo Kanade, Shohei Nobuhara, Yaser Sheikh
  • Action Detection by Implicit Intentional Motion Clustering [pdf] - Wei Chen, Jason J. Corso
  • Motion Trajectory Segmentation via Minimum Cost Multicuts [pdf] - Margret Keuper, Bjoern Andres, Thomas Brox
A Matrix Decomposition Perspective to Multiple Graph Matching [pdf]
Junchi Yan, Hongteng Xu, Hongyuan Zha, Xiaokang Yang, Huanxi Liu, Stephen Chu

Abstract: Graph matching has a wide spectrum of applications and in its general form is known NP-hard. In many vision tasks, one realistic problem arises for finding the global node mappings across a batch of corrupted weighted graphs. This paper is an attempt to connect graph matching, especially multi-graph matching to the matrix decomposition model and its relevant convex optimization algorithms. Our method aims to extract the common inliers and their synchronized permutations from disordered weighted graphs in the presence of deformation and outliers. Under the proposed framework, several variants can be derived in the hope of accommodating to specific types of noises. Experimental results empirically show that the proposed paradigm exhibits interesting behaviors and often performs competitively with the state-of-the-arts.
Similar papers:
  • Attribute-Graph: A Graph Based Approach to Image Ranking [pdf] - Nikita Prabhu, R. Venkatesh Babu
  • 3D Fragment Reassembly Using Integrated Template Guidance and Fracture-Region Matching [pdf] - Kang Zhang, Wuyi Yu, Mary Manhein, Warren Waggenspack, Xin Li
  • Multi-Image Matching via Fast Alternating Minimization [pdf] - Xiaowei Zhou, Menglong Zhu, Kostas Daniilidis
  • Mining And-Or Graphs for Graph Matching and Object Discovery [pdf] - Quanshi Zhang, Ying Nian Wu, Song-Chun Zhu
  • Discrete Tabu Search for Graph Matching [pdf] - Kamil Adamczewski, Yumin Suh, Kyoung Mu Lee
HD-CNN: Hierarchical Deep Convolutional Neural Networks for Large Scale Visual Recognition [pdf]
Zhicheng Yan, Hao Zhang, Robinson Piramuthu, Vignesh Jagadeesh, Dennis DeCoste, Wei Di, Yizhou Yu

Abstract: In image classification, visual separability between different object categories is highly uneven, and some categories are more difficult to distinguish than others. Such difficult categories demand more dedicated classifiers. However, existing deep convolutional neural networks (CNN) are trained as flat N-way classifiers, and few efforts have been made to leverage the hierarchical structure of categories. In this paper, we introduce hierarchical deep CNNs (HD-CNNs) by embedding deep CNNs into a two-level category hierarchy. An HD-CNN separates easy classes using a coarse category classifier while distinguishing difficult classes using fine category classifiers. During HDCNN training, component-wise pretraining is followed by global fine-tuning with a multinomial logistic loss regularized by a coarse category consistency term. In addition, conditional executions of fine category classifiers and layer parameter compression make HD-CNNs scalable for largescale visual recognition. We achieve state-of-the-art results on both CIFAR100 and large-scale ImageNet 1000-class benchmark datasets. In our experiments, we build up three different two-level HD-CNNs, and they lower the top-1 error of the standard CNNs by 2.65%, 3.1%, and 1.1%.
Similar papers:
  • What Makes an Object Memorable? [pdf] - Rachit Dubey, Joshua Peterson, Aditya Khosla, Ming-Hsuan Yang, Bernard Ghanem
  • Spatial Semantic Regularisation for Large Scale Object Detection [pdf] - Damian Mrowca, Marcus Rohrbach, Judy Hoffman, Ronghang Hu, Kate Saenko, Trevor Darrell
  • A Multiscale Variable-Grouping Framework for MRF Energy Minimization [pdf] - Omer Meir, Meirav Galun, Stav Yagev, Ronen Basri, Irad Yavneh
  • Human Parsing With Contextualized Convolutional Neural Network [pdf] - Xiaodan Liang, Chunyan Xu, Xiaohui Shen, Jianchao Yang, Si Liu, Jinhui Tang, Liang Lin, Shuicheng Yan
  • Relaxing From Vocabulary: Robust Weakly-Supervised Deep Learning for Vocabulary-Free Image Tagging [pdf] - Jianlong Fu, Yue Wu, Tao Mei, Jinqiao Wang, Hanqing Lu, Yong Rui
Convolutional Channel Features [pdf]
Bin Yang, Junjie Yan, Zhen Lei, Stan Z. Li

Abstract: Deep learning methods are powerful tools but often suffer from expensive computation and limited flexibility. An alternative is to combine light-weight models with deep representations. As successful cases exist in several visual problems, a unified framework is absent. In this paper, we revisit two widely used approaches in computer vision, namely filtered channel features and Convolutional Neural Networks (CNN), and absorb merits from both by proposing an integrated method called Convolutional Channel Features (CCF). CCF transfers low-level features from pretrained CNN models to feed the boosting forest model. With the combination of CNN features and boosting forest, CCF benefits from the richer capacity in feature representation compared with channel features, as well as lower cost in computation and storage compared with end-to-end CNN methods. We show that CCF serves as a good way of tailoring pre-trained CNN models to diverse tasks without finetuning the whole network to each task by achieving stateof-the-art performances in pedestrian detection, face detection, edge detection and object proposal generation.
Similar papers:
  • DeepBox: Learning Objectness With Convolutional Networks [pdf] - Weicheng Kuo, Bharath Hariharan, Jitendra Malik
  • DeepProposal: Hunting Objects by Cascading Deep Convolutional Layers [pdf] - Amir Ghodrati, Ali Diba, Marco Pedersoli, Tinne Tuytelaars, Luc Van Gool
  • Learning Complexity-Aware Cascades for Deep Pedestrian Detection [pdf] - Zhaowei Cai, Mohammad Saberian, Nuno Vasconcelos
  • Object Detection Using Generalization and Efficiency Balanced Co-Occurrence Features [pdf] - Haoyu Ren, Ze-Nian Li
  • Deep Learning Strong Parts for Pedestrian Detection [pdf] - Yonglong Tian, Ping Luo, Xiaogang Wang, Xiaoou Tang
Deep Fried Convnets [pdf]
Zichao Yang, Marcin Moczulski, Misha Denil, Nando de Freitas, Alex Smola, Le Song, Ziyu Wang

Abstract: The fully- connected layers of deep convolutional neural networks typically contain over 90% of the network parameters. Reducing the number of parameters while preserving predictive performance is critically important for training big models in distributed systems and for deployment in embedded devices. In this paper, we introduce a novel Adaptive Fastfood transform to reparameterize the matrix-vector multiplication of fully connected layers. Reparameterizing a fully connected layer with d inputs and n outputs with the Adaptive Fastfood transform reduces the storage and computational costs costs from O(nd) to O(n) and O(n log d) respectively. Using the Adaptive Fastfood transform in convolutional networks results in what we call a deep fried convnet. These convnets are end-to-end trainable, and enable us to attain substantial reductions in the number of parameters without affecting prediction accuracy on the MNIST and ImageNet datasets.
Similar papers:
  • Learning Spatiotemporal Features With 3D Convolutional Networks [pdf] - Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, Manohar Paluri
  • Matrix Backpropagation for Deep Networks With Structured Layers [pdf] - Catalin Ionescu, Orestis Vantzos, Cristian Sminchisescu
  • Learning The Structure of Deep Convolutional Networks [pdf] - Jiashi Feng, Trevor Darrell
  • Fast R-CNN [pdf] - Ross Girshick
  • An Exploration of Parameter Redundancy in Deep Networks With Circulant Projections [pdf] - Yu Cheng, Felix X. Yu, Rogerio S. Feris, Sanjiv Kumar, Alok Choudhary, Shi-Fu Chang
From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf]
Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang

Abstract: In this paper, we propose a novel deep convolutional network (DCN) that achieves outstanding performance on FDDB, PASCAL Face, and AFW. Specifically, our method achieves a high recall rate of 90.99% on the challenging FDDB benchmark, outperforming the state-of-the-art method [23] by a large margin of 2.91%. Importantly, we consider finding faces from a new perspective through scoring facial parts responses by their spatial structure and arrangement. The scoring mechanism is carefully formulated considering challenging cases where faces are only partially visible. This consideration allows our network to detect faces under severe occlusion and unconstrained pose variation, which are the main difficulty and bottleneck of most existing face detection approaches. We show that despite the use of DCN, our network can achieve practical runtime speed.
Similar papers:
  • Learning to Predict Saliency on Face Images [pdf] - Mai Xu, Yun Ren, Zulin Wang
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • Personalized Age Progression With Aging Dictionary [pdf] - Xiangbo Shu, Jinhui Tang, Hanjiang Lai, Luoqi Liu, Shuicheng Yan
  • Simultaneous Local Binary Feature Learning and Encoding for Face Recognition [pdf] - Jiwen Lu, Venice Erin Liong, Jie Zhou
  • Deep Learning Face Attributes in the Wild [pdf] - Ziwei Liu, Ping Luo, Xiaogang Wang, Xiaoou Tang
Multi-Scale Recognition With DAG-CNNs [pdf]
Songfan Yang, Deva Ramanan

Abstract: We explore multi-scale convolutional neural nets (CNNs) for image classification. Contemporary approaches extract features from a single output layer. By extracting features from multiple layers, one can simultaneously reason about high, mid, and low-level features during classification. The resulting multi-scale architecture can itself be seen as a feed-forward model that is structured as a directed acyclic graph (DAG-CNNs). We use DAG-CNNs to learn a set of multi-scale features that can be effectively shared between coarse and fine-grained classification tasks. While finetuning such models helps performance, we show that even off-the-self multi-scale features perform quite well. We present extensive analysis and demonstrate state-of-the-art classification performance on three standard scene benchmarks (SUN397, MIT67, and Scene15). In terms of the heavily benchmarked MIT67 and Scene15 datasets, our results reduce the lowest previously-reported error by 23.9% and 9.5%, respectively.
Similar papers:
  • Task-Driven Feature Pooling for Image Classification [pdf] - Guo-Sen Xie, Xu-Yao Zhang, Xiangbo Shu, Shuicheng Yan, Cheng-Lin Liu
  • Learning The Structure of Deep Convolutional Networks [pdf] - Jiashi Feng, Trevor Darrell
  • Fast R-CNN [pdf] - Ross Girshick
  • Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification [pdf] - Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun
  • Matrix Backpropagation for Deep Networks With Structured Layers [pdf] - Catalin Ionescu, Orestis Vantzos, Cristian Sminchisescu
Resolving Scale Ambiguity Via XSlit Aspect Ratio Analysis [pdf]
Wei Yang, Haiting Lin, Sing Bing Kang, Jingyi Yu

Abstract: In perspective cameras, images of a frontal-parallel 3D object preserve its aspect ratio invariant to its depth. Such an invariance is useful in photography but is unique to perspective projection. In this paper, we show that alternative non-perspective cameras such as the crossed-slit or XSlit cameras exhibit a different depth-dependent aspect ratio (DDAR) property that can be used to 3D recovery. We first conduct a comprehensive analysis to characterize DDAR, infer object depth from its AR, and model recoverable depth range, sensitivity, and error. We show that repeated shape patterns in real Manhattan World scenes can be used for 3D reconstruction using a single XSlit image. We also extend our analysis to model slopes of lines. Specifically, parallel 3D lines exhibit depth-dependent slopes (DDS) on their images which can also be used to infer their depths. We validate our analyses using real XSlit cameras, XSlit panoramas, and catadioptric mirrors. Experiments show that DDAR and DDS provide important depth cues and enable effective single-image scene reconstruction.
Similar papers:
  • Global Structure-From-Motion by Similarity Averaging [pdf] - Zhaopeng Cui, Ping Tan
  • Model-Based Tracking at 300Hz Using Raw Time-of-Flight Observations [pdf] - Jan Stuhmer, Sebastian Nowozin, Andrew Fitzgibbon, Richard Szeliski, Travis Perry, Sunil Acharya, Daniel Cremers, Jamie Shotton
  • Occlusion-Aware Depth Estimation Using Light-Field Cameras [pdf] - Ting-Chun Wang, Alexei A. Efros, Ravi Ramamoorthi
  • Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf] - Naejin Kong, Michael J. Black
  • High Quality Structure From Small Motion for Rolling Shutter Cameras [pdf] - Sunghoon Im, Hyowon Ha, Gyeongmin Choe, Hae-Gon Jeon, Kyungdon Joo, In So Kweon
Self-Occlusions and Disocclusions in Causal Video Object Segmentation [pdf]
Yanchao Yang, Ganesh Sundaramoorthi, Stefano Soatto

Abstract: We propose a method to detect disocclusion in video sequences of three-dimensional scenes and to partition the disoccluded regions into objects, defined by coherent deformation corresponding to surfaces in the scene. Our method infers deformation fields that are piecewise smooth by construction without the need for an explicit regularizer and the associated choice of weight. It then partitions the disoccluded region and groups its components with objects by leveraging on the complementarity of motion and appearance cues: Where appearance changes within an object, motion can usually be reliably inferred and used for grouping. Where appearance is close to constant, it can be used for grouping directly. We integrate both cues in an energy minimization framework, incorporate prior assumptions explicitly into the energy, and propose a numerical scheme.
Similar papers:
  • Unsupervised Object Discovery and Tracking in Video Collections [pdf] - Suha Kwak, Minsu Cho, Ivan Laptev, Jean Ponce, Cordelia Schmid
  • Robust Non-Rigid Motion Tracking and Surface Reconstruction Using L0 Regularization [pdf] - Kaiwen Guo, Feng Xu, Yangang Wang, Yebin Liu, Qionghai Dai
  • Efficient Video Segmentation Using Parametric Graph Partitioning [pdf] - Chen-Ping Yu, Hieu Le, Gregory Zelinsky, Dimitris Samaras
  • Motion Trajectory Segmentation via Minimum Cost Multicuts [pdf] - Margret Keuper, Bjoern Andres, Thomas Brox
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
Unsupervised Extraction of Video Highlights Via Robust Recurrent Auto-Encoders [pdf]
Huan Yang, Baoyuan Wang, Stephen Lin, David Wipf, Minyi Guo, Baining Guo

Abstract: With the growing popularity of short-form video sharing platforms such as Instagram and Vine, there has been an increasing need for techniques that automatically extract highlights from video. Whereas prior works have approached this problem with heuristic rules or supervised learning, we present an unsupervised learning approach that takes advantage of the abundance of user-edited videos on social media websites such as YouTube. Based on the idea that the most significant sub-events within a video class are commonly present among edited videos while less interesting ones appear less frequently, we identify the signifi- cant sub-events via a robust recurrent auto-encoder trained on a collection of user-edited videos queried for each particular class of interest. The auto-encoder is trained using a proposed shrinking exponential loss function that makes it robust to noise in the web-crawled training data, and is configured with bidirectional long short term memory (LSTM) [5] cells to better model the temporal structure of highlight segments. Different from supervised techniques, our method can infer highlights using only a set of downloaded edited videos, without also needing their pre-edited counterparts which are rarely available online. Extensive experiments indicate the promise of our proposed solution in this challenging unsupervised setting.
Similar papers:
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
  • Sequence to Sequence - Video to Text [pdf] - Subhashini Venugopalan, Marcus Rohrbach, Jeffrey Donahue, Raymond Mooney, Trevor Darrell, Kate Saenko
  • Semantic Video Entity Linking Based on Visual Content and Metadata [pdf] - Yuncheng Li, Xitong Yang, Jiebo Luo
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
Describing Videos by Exploiting Temporal Structure [pdf]
Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville

Abstract: Recent progress in using recurrent neural networks (RNNs) for image description has motivated the exploration of their application for video description. However, while images are static, working with videos requires modeling their dynamic temporal structure and then properly integrating that information into a natural language description. In this context, we propose an approach that successfully takes into account both the local and global temporal structure of videos to produce descriptions. First, our approach incorporates a spatial temporal 3-D convolutional neural network (3-D CNN) representation of the short temporal dynamics. The 3-D CNN representation is trained on video action recognition tasks, so as to produce a representation that is tuned to human motion and behavior. Second we propose a temporal attention mechanism that allows to go beyond local temporal modeling and learns to automatically select the most relevant temporal segments given the text-generating RNN. Our approach exceeds the current state-of-art for both BLEU and METEOR metrics on the Youtube2Text dataset. We also present results on a new, larger and more challenging dataset of paired video and natural language descriptions.
Similar papers:
  • Learning Spatiotemporal Features With 3D Convolutional Networks [pdf] - Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, Manohar Paluri
  • Temporal Perception and Prediction in Ego-Centric Video [pdf] - Yipin Zhou, Tamara L. Berg
  • Human Action Recognition Using Factorized Spatio-Temporal Convolutional Networks [pdf] - Lin Sun, Kui Jia, Dit-Yan Yeung, Bertram E. Shi
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
  • Sequence to Sequence - Video to Text [pdf] - Subhashini Venugopalan, Marcus Rohrbach, Jeffrey Donahue, Raymond Mooney, Trevor Darrell, Kate Saenko
Learning Query and Image Similarities With Ranking Canonical Correlation Analysis [pdf]
Ting Yao, Tao Mei, Chong-Wah Ngo

Abstract: One of the fundamental problems in image search is to learn the ranking functions, i.e., similarity between the query and image. The research on this topic has evolved through two paradigms: feature-based vector model and image ranker learning. The former relies on the image surrounding texts, while the latter learns a ranker based on human labeled query-image pairs. Each of the paradigms has its own limitation. The vector model is sensitive to the quality of text descriptions, and the learning paradigm is difficult to be scaled up as human labeling is always too expensive to obtain. We demonstrate in this paper that the above two limitations can be well mitigated by jointly exploring subspace learning and the use of click-through data. Specifically, we propose a novel Ranking Canonical Correlation Analysis (RCCA) for learning query and image similarities. RCCA initially finds a common subspace between query and image views by maximizing their correlations, and further simultaneously learns a bilinear queryimage similarity function and adjusts the subspace to preserve the preference relations implicit in the click-through data. Once the subspace is finalized, query-image similarity can be computed by the bilinear similarity function on their mappings in this subspace. On a large-scale click-based image dataset with 11.7 million queries and one million images, RCCA is shown to be powerful for image search with superior performance over several state-of-the-art methods on both
Similar papers:
  • Multi-View Subspace Clustering [pdf] - Hongchang Gao, Feiping Nie, Xuelong Li, Heng Huang
  • Query Adaptive Similarity Measure for RGB-D Object Recognition [pdf] - Yanhua Cheng, Rui Cai, Chi Zhang, Zhiwei Li, Xin Zhao, Kaiqi Huang, Yong Rui
  • Similarity Gaussian Process Latent Variable Model for Multi-Modal Data Analysis [pdf] - Guoli Song, Shuhui Wang, Qingming Huang, Qi Tian
  • Scalable Person Re-Identification: A Benchmark [pdf] - Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian
  • Attribute-Graph: A Graph Based Approach to Image Ranking [pdf] - Nikita Prabhu, R. Venkatesh Babu
Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf]
Saehoon Yi, Vladimir Pavlovic

Abstract: Video segmentation is a stepping stone to understanding video context. Video segmentation enables one to represent a video by decomposing it into coherent regions which comprise whole or parts of objects. However, the challenge originates from the fact that most of the video segmentation algorithms are based on unsupervised learning due to expensive cost of pixelwise video annotation and intra-class variability within similar unconstrained video classes. We propose a Markov Random Field model for unconstrained video segmentation that relies on tight integration of multiple cues: vertices are defined from contour based superpixels, unary potentials from temporally smooth label likelihood and pairwise potentials from global structure of a video. Multi-cue structure is a breakthrough to extracting coherent object regions for unconstrained videos in absence of supervision. Our experiments on VSB100 dataset show that the proposed model significantly outperforms competing state-of-the-art algorithms. Qualitative analysis illustrates that video segmentation result of the proposed model is consistent with human perception of objects.
Similar papers:
  • Multiresolution Hierarchy Co-Clustering for Semantic Segmentation in Sequences With Small Variations [pdf] - David Varas, Monica Alfaro, Ferran Marques
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
  • Fully Connected Object Proposals for Video Segmentation [pdf] - Federico Perazzi, Oliver Wang, Markus Gross, Alexander Sorkine-Hornung
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
  • Efficient Video Segmentation Using Parametric Graph Partitioning [pdf] - Chen-Ping Yu, Hieu Le, Gregory Zelinsky, Dimitris Samaras
Pedestrian Travel Time Estimation in Crowded Scenes [pdf]
Shuai Yi, Hongsheng Li, Xiaogang Wang

Abstract: In this paper, we target on the problem of estimating the statistic of pedestrian travel time within a period from an entrance to a destination in a crowded scene. Such estimation is based on the global distributions of crowd densities and velocities instead of complete trajectories of pedestrians, which cannot be obtained in crowded scenes. The proposed method is motivated by our statistical investigation into the correlations between travel time and global properties of crowded scenes. Active regions are created for each source-destination pair to model the probable walking regions over the corresponding source-destination traffic flow. Two sets of scene features are specially designed for modeling moving and stationary persons inside the active regions and their influences on pedestrian travel time. The estimation of pedestrian travel time provides valuable information for both crowd scene understanding and pedestrian behavior analysis, but was not sufficiently studied in literature. The effectiveness of the proposed pedestrian travel time estimation model is demonstrated through several surveillance applications, including dynamic scene monitoring, localization of regions blocking traffics, and detection of abnormal pedestrian behaviors. Many more valuable applications based on our method are to be explored in the future.
Similar papers:
  • Active Transfer Learning With Zero-Shot Priors: Reusing Past Datasets for Future Tasks [pdf] - Efstratios Gavves, Thomas Mensink, Tatiana Tommasi, Cees G. M. Snoek, Tinne Tuytelaars
  • Convolutional Channel Features [pdf] - Bin Yang, Junjie Yan, Zhen Lei, Stan Z. Li
  • COUNT Forest: CO-Voting Uncertain Number of Targets Using Random Forest for Crowd Density Estimation [pdf] - Viet-Quoc Pham, Tatsuo Kozakaya, Osamu Yamaguchi, Ryuzo Okada
  • A Spatio-Temporal Appearance Representation for Viceo-Based Pedestrian Re-Identification [pdf] - Kan Liu, Bingpeng Ma, Wei Zhang, Rui Huang
  • Deep Learning Strong Parts for Pedestrian Detection [pdf] - Yonglong Tian, Ping Luo, Xiaogang Wang, Xiaoou Tang
Extended Depth of Field Catadioptric Imaging Using Focal Sweep [pdf]
Ryunosuke Yokoya, Shree K. Nayar

Abstract: Catadioptric imaging systems use curved mirrors to capture wide fields of view. However, due to the curvature of the mirror, these systems tend to have very limited depth of field (DOF), with the point spread function (PSF) varying dramatically over the field of view and as a function of scene depth. In recent years, focal sweep has been used extensively to extend the DOF of conventional imaging systems. It has been shown that focal sweep produces an integrated point spread function (IPSF) that is nearly space-invariant and depth-invariant, enabling the recovery of an extended depth of field (EDOF) image by deconvolving the captured focal sweep image with a single IPSF. In this paper, we use focal sweep to extend the DOF of a catadioptric imaging system. We show that while the IPSF is spatially varying when a curved mirror is used, it remains quasi depthinvariant over the wide field of view of the imaging system. We have developed a focal sweep system where mirrors of different shapes can be used to capture wide field of view EDOF images. In particular, we show experimental results using spherical and paraboloidal mirrors.
Similar papers:
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • On the Equivalence of Moving Entrance Pupil and Radial Distortion for Camera Calibration [pdf] - Avinash Kumar, Narendra Ahuja
  • Amodal Completion and Size Constancy in Natural Scenes [pdf] - Abhishek Kar, Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf] - Kensei Jo, Mohit Gupta, Shree K. Nayar
  • Depth Recovery From Light Field Using Focal Stack Symmetry [pdf] - Haiting Lin, Can Chen, Sing Bing Kang, Jingyi Yu
AttentionNet: Aggregating Weak Directions for Accurate Object Detection [pdf]
Donggeun Yoo, Sunggyun Park, Joon-Young Lee, Anthony S. Paek, In So Kweon

Abstract: We present a novel detection method using a deep convolutional neural network (CNN), named AttentionNet. We cast an object detection problem as an iterative classification problem, which is the most suitable form of a CNN. AttentionNet provides quantized weak directions pointing a target object and the ensemble of iterative predictions from AttentionNet converges to an accurate object boundary box. Since AttentionNet is a unified network for object detection, it detects objects without any separated models from the object proposal to the post bounding-box regression. We evaluate AttentionNet by a human detection task and achieve the state-of-the-art performance of 65% (AP) on PASCAL VOC 2007/2012 with an 8-layered architecture only.
Similar papers:
  • Weakly- and Semi-Supervised Learning of a Deep Convolutional Network for Semantic Image Segmentation [pdf] - George Papandreou, Liang-Chieh Chen, Kevin P. Murphy, Alan L. Yuille
  • Object Detection via a Multi-Region and Semantic Segmentation-Aware CNN Model [pdf] - Spyros Gidaris, Nikos Komodakis
  • Cascaded Sparse Spatial Bins for Efficient and Effective Generic Object Detection [pdf] - David Novotny, Jiri Matas
  • Amodal Completion and Size Constancy in Natural Scenes [pdf] - Abhishek Kar, Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • Oriented Object Proposals [pdf] - Shengfeng He, Rynson W.H. Lau
Direct, Dense, and Deformable: Template-Based Non-Rigid 3D Reconstruction From RGB Video [pdf]
Rui Yu, Chris Russell, Neill D. F. Campbell, Lourdes Agapito

Abstract: In this paper we tackle the problem of capturing the dense, detailed 3D geometry of generic, complex non-rigid meshes using a single RGB-only commodity video camera and a direct approach. While robust and even real-time solutions exist to this problem if the observed scene is static, for non-rigid dense shape capture current systems are typically restricted to the use of complex multi-camera rigs, take advantage of the additional depth channel available in RGB-D cameras, or deal with specific shapes such as faces or planar surfaces. In contrast, our method makes use of a single RGB video as input; it can capture the deformations of generic shapes; and the depth estimation is dense, per-pixel and direct. We first compute a dense 3D template of the shape of the object, using a short rigid sequence, and subsequently perform online reconstruction of the non-rigid mesh as it evolves over time. Our energy optimization approach minimizes a robust photometric cost that simultaneously estimates the temporal correspondences and 3D deformations with respect to the template mesh. In our experimental evaluation we show a range of qualitative results on novel datasets; we compare against an existing method that requires multi-frame optical flow; and perform a quantitative evaluation against other template-based approaches on a ground truth dataset.
Similar papers:
  • Deformable 3D Fusion: From Partial Dynamic 3D Observations to Complete 4D Models [pdf] - Weipeng Xu, Mathieu Salzmann, Yongtian Wang, Yue Liu
  • Detailed Full-Body Reconstructions of Moving People From Monocular RGB-D Sequences [pdf] - Federica Bogo, Michael J. Black, Matthew Loper, Javier Romero
  • Exploiting Object Similarity in 3D Reconstruction [pdf] - Chen Zhou, Fatma Guney, Yizhou Wang, Andreas Geiger
  • Face Flow [pdf] - Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou
  • Dense Image Registration and Deformable Surface Reconstruction in Presence of Occlusions and Minimal Texture [pdf] - Dat Tien Ngo, Sanghyuk Park, Anne Jorstad, Alberto Crivellaro, Chang D. Yoo, Pascal Fua
Efficient Video Segmentation Using Parametric Graph Partitioning [pdf]
Chen-Ping Yu, Hieu Le, Gregory Zelinsky, Dimitris Samaras

Abstract: Video segmentation is the task of grouping similar pixels in the spatio-temporal domain, and has become an important preprocessing step for subsequent video analysis. Most video segmentation and supervoxel methods output a hierarchy of segmentations, but while this provides useful multiscale information, it also adds difficulty in selecting the appropriate level for a task. In this work, we propose an efficient and robust video segmentation framework based on parametric graph partitioning (PGP), a fast, almost parameter free graph partitioning method that identifies and removes between-cluster edges to form node clusters. Apart from its computational efficiency, PGP performs clustering of the spatio-temporal volume without requiring a pre-specified cluster number or bandwidth parameters, thus making video segmentation more practical to use in applications. The PGP framework also allows processing sub-volumes, which further improves performance, contrary to other streaming video segmentation methods where sub-volume processing reduces performance. We evaluate the PGP method using the SegTrack v2 and Chen Xiph.org datasets, and show that it outperforms related state-of-theart algorithms in 3D segmentation metrics and running time.
Similar papers:
  • Self-Occlusions and Disocclusions in Causal Video Object Segmentation [pdf] - Yanchao Yang, Ganesh Sundaramoorthi, Stefano Soatto
  • Motion Trajectory Segmentation via Minimum Cost Multicuts [pdf] - Margret Keuper, Bjoern Andres, Thomas Brox
  • Video Segmentation With Just a Few Strokes [pdf] - Naveen Shankar Nagaraja, Frank R. Schmidt, Thomas Brox
  • Fully Connected Object Proposals for Video Segmentation [pdf] - Federico Perazzi, Oliver Wang, Markus Gross, Alexander Sorkine-Hornung
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
Fill and Transfer: A Simple Physics-Based Approach for Containability Reasoning [pdf]
Lap-Fai Yu, Noah Duncan, Sai-Kit Yeung

Abstract: The visual perception of object affordances has emerged as a useful ingredient for building powerful computer vision and robotic applications [31]. In this paper we introduce a novel approach to reason about liquid containability - the affordance of containing liquid. Our approach analyzes container objects based on two simple physical processes: the Fill and Transfer of liquid. First, it reasons about whether a given 3D object is a liquid container and its best filling direction. Second, it proposes directions to transfer its contained liquid to the outside while avoiding spillage. We compare our simplified model with a common fluid dynamics simulation and demonstrate that our algorithm makes human-like choices about the best directions to fill containers and transfer liquid from them. We apply our approach to reason about the containability of several real-world objects acquired using a consumer-grade depth camera.
Similar papers:
  • Variational PatchMatch MultiView Reconstruction and Refinement [pdf] - Philipp Heise, Brian Jensen, Sebastian Klose, Alois Knoll
  • Single-Shot Specular Surface Reconstruction With Gonio-Plenoptic Imaging [pdf] - Lingfei Meng, Liyang Lu, Noah Bedard, Kathrin Berkner
  • Bayesian Model Adaptation for Crowd Counts [pdf] - Bo Liu, Nuno Vasconcelos
  • Amodal Completion and Size Constancy in Natural Scenes [pdf] - Abhishek Kar, Shubham Tulsiani, Joao Carreira, Jitendra Malik
  • Compression Artifacts Reduction by a Deep Convolutional Network [pdf] - Chao Dong, Yubin Deng, Chen Change Loy, Xiaoou Tang
Just Noticeable Differences in Visual Attributes [pdf]
Aron Yu, Kristen Grauman

Abstract: We explore the problem of predicting just noticeable differences in a visual attribute. While some pairs of images have a clear ordering for an attribute (e.g., A is more sporty than B), for others the difference may be indistinguishable to human observers. However, existing relative attribute models are unequipped to infer partial orders on novel data. Attempting to map relative attribute ranks to equality predictions is non-trivial, particularly since the span of indistinguishable pairs in attribute space may vary in different parts of the feature space. We develop a Bayesian local learning strategy to infer when images are indistinguishable for a given attribute. On the UT-Zap50K shoes and LFW-10 faces datasets, we outperform a variety of alternative methods. In addition, we show the practical impact on fine-grained visual search.
Similar papers:
  • Attributed Grammars for Joint Estimation of Human Attributes, Part and Pose [pdf] - Seyoung Park, Song-Chun Zhu
  • Multi-Task Learning With Low Rank Attribute Embedding for Person Re-Identification [pdf] - Chi Su, Fan Yang, Shiliang Zhang, Qi Tian, Larry S. Davis, Wen Gao
  • Learning Deep Representation With Large-Scale Attributes [pdf] - Wanli Ouyang, Hongyang Li, Xingyu Zeng, Xiaogang Wang
  • Discovering the Spatial Extent of Relative Attributes [pdf] - Fanyi Xiao, Yong Jae Lee
  • A Unified Multiplicative Framework for Attribute Learning [pdf] - Kongming Liang, Hong Chang, Shiguang Shan, Xilin Chen
Piecewise Flat Embedding for Image Segmentation [pdf]
Yizhou Yu, Chaowei Fang, Zicheng Liao

Abstract: Image segmentation is a critical step in many computer vision tasks, including high-level visual recognition and scene understanding as well as low-level photo and video processing. In this paper, we propose a new nonlinear embedding, called piecewise flat embedding, for image segmentation. Based on the theory of sparse signal recovery, piecewise flat embedding attempts to identify segment boundaries while significantly suppressing variations within segments. We adopt an L1 -regularized energy term in the formulation to promote sparse solutions. We further devise an effective two-stage numerical algorithm based on Bregman iterations to solve the proposed embedding. Piecewise flat embedding can be easily integrated into existing image segmentation frameworks, including segmentation based on spectral clustering and hierarchical segmentation based on contour detection. Experiments on BSDS500 indicate that segmentation algorithms incorporating this embedding can achieve significantly improved results in both frameworks.
Similar papers:
  • Multi-Cue Structure Preserving MRF for Unconstrained Video Segmentation [pdf] - Saehoon Yi, Vladimir Pavlovic
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
  • Secrets of GrabCut and Kernel K-Means [pdf] - Meng Tang, Ismail Ben Ayed, Dmitrii Marin, Yuri Boykov
  • Classical Scaling Revisited [pdf] - Gil Shamai, Yonathan Aflalo, Michael Zibulevsky, Ron Kimmel
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
Visual Madlibs: Fill in the Blank Description Generation and Question Answering [pdf]
Licheng Yu, Eunbyung Park, Alexander C. Berg, Tamara L. Berg

Abstract: In this paper, we introduce a new dataset consisting of 360,001 focused natural language descriptions for 10,738 images. This dataset, the Visual Madlibs dataset, is collected using automatically produced fill-in-the-blank templates designed to gather targeted descriptions about: people and objects, their appearances, activities, and interactions, as well as inferences about the general scene or its broader context. We provide several analyses of the Visual Madlibs dataset and demonstrate its applicability to two new description generation tasks: focused description generation, and multiple-choice question-answering for images. Experiments using joint-embedding and deep learning methods show promising results on these tasks.
Similar papers:
  • Guiding the Long-Short Term Memory Model for Image Caption Generation [pdf] - Xu Jia, Efstratios Gavves, Basura Fernando, Tinne Tuytelaars
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
  • Sequence to Sequence - Video to Text [pdf] - Subhashini Venugopalan, Marcus Rohrbach, Jeffrey Donahue, Raymond Mooney, Trevor Darrell, Kate Saenko
  • Ask Your Neurons: A Neural-Based Approach to Answering Questions About Images [pdf] - Mateusz Malinowski, Marcus Rohrbach, Mario Fritz
  • VQA: Visual Question Answering [pdf] - Stanislaw Antol, Aishwarya Agrawal, Jiasen Lu, Margaret Mitchell, Dhruv Batra, C. Lawrence Zitnick, Devi Parikh
Large Displacement 3D Scene Flow With Occlusion Reasoning [pdf]
Andrei Zanfir, Cristian Sminchisescu

Abstract: The emergence of modern, affordable and accurate RGB-D sensors increases the need for single view approaches to estimate 3-dimensional motion, also known as scene flow. In this paper we propose a coarse-to-fine, dense, correspondence-based scene flow formulation that relies on explicit geometric reasoning to account for the effects of large displacements and to model occlusion. Our methodology enforces local motion rigidity at the level of the 3d point cloud without explicitly smoothing the parameters of adjacent neighborhoods. By integrating all geometric and photometric components in a single, consistent, occlusion-aware energy model, defined over overlapping, image-adaptive neighborhoods, our method can process fast motions and large occlusions areas, as present in challenging datasets like the MPI Sintel Flow Dataset, recently augmented with depth information. By explicitly modeling large displacements and occlusion, we can handle difficult sequences which cannot be currently processed by state of the art scene flow methods. We also show that by integrating depth information into the model, we can obtain correspondence fields with improved spatial support and sharper boundaries compared to the state of the art, large-displacement optical flow methods.
Similar papers:
  • Photogeometric Scene Flow for High-Detail Dynamic 3D Reconstruction [pdf] - Paulo F. U. Gotardo, Tomas Simon, Yaser Sheikh, Iain Matthews
  • Face Flow [pdf] - Patrick Snape, Anastasios Roussos, Yannis Panagakis, Stefanos Zafeiriou
  • Oriented Light-Field Windows for Scene Flow [pdf] - Pratul P. Srinivasan, Michael W. Tao, Ren Ng, Ravi Ramamoorthi
  • FlowNet: Learning Optical Flow With Convolutional Networks [pdf] - Alexey Dosovitskiy, Philipp Fischer, Eddy Ilg, Philip Hausser, Caner Hazirbas, Vladimir Golkov, Patrick van der Smagt, Daniel Cremers, Thomas Brox
  • Flow Fields: Dense Correspondence Fields for Highly Accurate Large Displacement Optical Flow Estimation [pdf] - Christian Bailer, Bertram Taetz, Didier Stricker
Camera Pose Voting for Large-Scale Image-Based Localization [pdf]
Bernhard Zeisl, Torsten Sattler, Marc Pollefeys

Abstract: Image-based localization approaches aim to determine the camera pose from which an image was taken. Finding correct 2D-3D correspondences between query image features and 3D points in the scene model becomes harder as the size of the model increases. Current state-of-theart methods therefore combine elaborate matching schemes with camera pose estimation techniques that are able to handle large fractions of wrong matches. In this work we study the benefits and limitations of spatial verification compared to appearance-based filtering. We propose a voting-based pose estimation strategy that exhibits O(n) complexity in the number of matches and thus facilitates to consider much more matches than previous approaches whose complexity grows at least quadratically. This new outlier rejection formulation enables us to evaluate pose estimation for 1-to-many matches and to surpass the state-ofthe-art. At the same time, we show that using more matches does not automatically lead to a better performance.
Similar papers:
  • PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization [pdf] - Alex Kendall, Matthew Grimes, Roberto Cipolla
  • Learning Where to Position Parts in 3D [pdf] - Marco Pedersoli, Tinne Tuytelaars
  • Adaptive Dither Voting for Robust Spatial Verification [pdf] - Xiaomeng Wu, Kunio Kashino
  • Localize Me Anywhere, Anytime: A Multi-Task Point-Retrieval Approach [pdf] - Guoyu Lu, Yan Yan, Li Ren, Jingkuan Song, Nicu Sebe, Chandra Kambhamettu
  • Hyperpoints and Fine Vocabularies for Large-Scale Location Recognition [pdf] - Torsten Sattler, Michal Havlena, Filip Radenovic, Konrad Schindler, Marc Pollefeys
CV-HAZOP: Introducing Test Data Validation for Computer Vision [pdf]
Oliver Zendel, Markus Murschitz, Martin Humenberger, Wolfgang Herzner

Abstract: Test data plays an important role in computer vision (CV) but is plagued by two questions: Which situations should be covered by the test data and have we tested enough to reach a conclusion? In this paper we propose a new solution answering these questions using a standard procedure devised by the safety community to validate complex systems: The Hazard and Operability Analysis (HAZOP). It is designed to systematically search and identify difficult, performance-decreasing situations and aspects. We introduce a generic CV model that creates the basis for the hazard analysis and, for the first time, apply an extensive HAZOP to the CV domain. The result is a publicly available checklist with more than 900 identified individual hazards. This checklist can be used to evaluate existing test datasets by quantifying the amount of covered hazards. We evaluate our approach by first analyzing and annotating the popular stereo vision test datasets Middlebury and KITTI. Second, we compare the performance of six popular stereo matching algorithms at the identified hazards from our checklist with their average performance and show, as expected, a clear negative influence of the hazards. The presented approach is a useful tool to evaluate and improve test datasets and creates a common basis for future dataset designs. This work presents a new way to facilitate this safety assessment process and goes beyond the basic scope of benchmarking: applying a proven method used by the safety communi
Similar papers:
  • A Deep Visual Correspondence Embedding Model for Stereo Matching Costs [pdf] - Zhuoyuan Chen, Xun Sun, Liang Wang, Yinan Yu, Chang Huang
  • StereoSnakes: Contour Based Consistent Object Extraction For Stereo Images [pdf] - Ran Ju, Tongwei Ren, Gangshan Wu
  • Photometric Stereo With Small Angular Variations [pdf] - Jian Wang, Yasuyuki Matsushita, Boxin Shi, Aswin C. Sankaranarayanan
  • The HCI Stereo Metrics: Geometry-Aware Performance Analysis of Stereo Algorithms [pdf] - Katrin Honauer, Lena Maier-Hein, Daniel Kondermann
  • Photometric Stereo in a Scattering Medium [pdf] - Zak Murez, Tali Treibitz, Ravi Ramamoorthi, David Kriegman
Confidence Preserving Machine for Facial Action Unit Detection [pdf]
Jiabei Zeng, Wen-Sheng Chu, Fernando De la Torre, Jeffrey F. Cohn, Zhang Xiong

Abstract: easy negative samples hard samples easy positive samples
Similar papers:
  • Bi-Shifting Auto-Encoder for Unsupervised Domain Adaptation [pdf] - Meina Kan, Shiguang Shan, Xilin Chen
  • Pairwise Conditional Random Forests for Facial Expression Recognition [pdf] - Arnaud Dapogny, Kevin Bailly, Severine Dubuisson
  • Learning to Transfer: Transferring Latent Task Structures and Its Application to Person-Specific Facial Action Unit Detection [pdf] - Timur Almaev, Brais Martinez, Michel Valstar
  • Multi-View Domain Generalization for Visual Recognition [pdf] - Li Niu, Wen Li, Dong Xu
  • From Emotions to Action Units With Hidden and Semi-Hidden-Task Learning [pdf] - Adria Ruiz, Joost Van de Weijer, Xavier Binefa
3D Fragment Reassembly Using Integrated Template Guidance and Fracture-Region Matching [pdf]
Kang Zhang, Wuyi Yu, Mary Manhein, Warren Waggenspack, Xin Li

Abstract: This paper studies matching of fragmented objects to recompose their original geometry. Solving this geometric reassembly problem has direct applications in archaeology and forensic investigation in the computer-aided restoration of damaged artifacts and evidence. We develop a new algorithm to effectively integrate both guidance from a template and from matching of adjacent pieces fractureregions. First, we compute partial matchings between fragments and a template, and pairwise matchings among fragments. Many potential matches are obtained and then selected/refined in a multi-piece matching stage to maximize global groupwise matching consistency. This pipeline is effective in composing fragmented thin-shell objects containing small pieces, whose pairwise matching is usually unreliable and ambiguous and hence their reassembly remains challenging to the existing algorithms. to predict the skulls ancestry and perform facial reconstruction/superimposition to help identify the body. The skeletal remains, however, are often incomplete or fragmented due to trauma or environmental exposure. Therefore, forensic specialists need to first perform a manual recomposition of skull fragments before subsequent assessment and analysis. With effective digital reassembly algorithms, after all the fragments are digitized, the existing manual skull reassembly procedure can be automated and augmented. Considering the geometry of a fragment, its boundary surface consists of two types of regions
Similar papers:
  • Mining And-Or Graphs for Graph Matching and Object Discovery [pdf] - Quanshi Zhang, Ying Nian Wu, Song-Chun Zhu
  • Multi-Image Matching via Fast Alternating Minimization [pdf] - Xiaowei Zhou, Menglong Zhu, Kostas Daniilidis
  • A Matrix Decomposition Perspective to Multiple Graph Matching [pdf] - Junchi Yan, Hongteng Xu, Hongyuan Zha, Xiaokang Yang, Huanxi Liu, Stephen Chu
  • Learning to Boost Filamentary Structure Segmentation [pdf] - Lin Gu, Li Cheng
  • Contour Guided Hierarchical Model for Shape Matching [pdf] - Yuanqi Su, Yuehu Liu, Bonan Cuan, Nanning Zheng
A Self-Paced Multiple-Instance Learning Framework for Co-Saliency Detection [pdf]
Dingwen Zhang, Deyu Meng, Chao Li, Lu Jiang, Qian Zhao, Junwei Han

Abstract: As an interesting and emerging topic, co-saliency detection aims at simultaneously extracting common salient objects in a group of images. Traditional co-saliency detection approaches heavily rely on human knowledge for designing hand-crafted metrics to explore the intrinsic patterns underlying co-salient objects. Such strategies, however, always suffer from poor generalization capability to flexibly adapt to various scenarios in real applications, especially due to their lack of insightful understanding of the biological mechanisms of human visual co-attention. To alleviate this problem, we propose a novel framework for this task, by naturally reformulating it as a multiple-instance learning (MIL) problem and further integrating it into a self-paced learning (SPL) regime. The proposed framework on one hand is capable of fitting insightful metric measurements and discovering common patterns under co-salient regions in a self-learning way by MIL, and on the other hand tends to promise the learning reliability and stability by simulating the human learning process through SPL. Experiments on benchmark datasets have demonstrated the effectiveness of the proposed framework as compared with the state-of-the-arts.1
Similar papers:
  • Towards Computational Baby Learning: A Weakly-Supervised Approach for Object Detection [pdf] - Xiaodan Liang, Si Liu, Yunchao Wei, Luoqi Liu, Liang Lin, Shuicheng Yan
  • SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf] - Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao
  • HARF: Hierarchy-Associated Rich Features for Salient Object Detection [pdf] - Wenbin Zou, Nikos Komodakis
  • Relaxed Multiple-Instance SVM With Application to Object Discovery [pdf] - Xinggang Wang, Zhuotun Zhu, Cong Yao, Xiang Bai
  • Simpler Non-Parametric Methods Provide as Good or Better Results to Multiple-Instance Learning [pdf] - Ragav Venkatesan, Parag Chandakkar, Baoxin Li
Fast Orthogonal Projection Based on Kronecker Product [pdf]
Xu Zhang, Felix X. Yu, Ruiqi Guo, Sanjiv Kumar, Shengjin Wang, Shi-Fu Chang

Abstract: We propose a family of structured matrices to speed up orthogonal projections for high-dimensional data commonly seen in computer vision applications. In this, a structured matrix is formed by the Kronecker product of a series of smaller orthogonal matrices. This achieves O(d log d) computational complexity and O(log d) space complexity for d-dimensional data, a drastic improvement over the standard unstructured projections whose computational and space complexities are both O(d2 ). We also introduce an efficient learning procedure for optimizing such matrices in a data dependent fashion. We demonstrate the significant advantages of the proposed approach in solving the approximate nearest neighbor (ANN) image search problem with both binary embedding and quantization. Comprehensive experiments show that the proposed approach can achieve similar or better accuracy as the existing state-of-the-art but with significantly less time and memory.
Similar papers:
  • An NMF Perspective on Binary Hashing [pdf] - Lopamudra Mukherjee, Sathya N. Ravi, Vamsi K. Ithapu, Tyler Holmes, Vikas Singh
  • Learning Binary Codes for Maximum Inner Product Search [pdf] - Fumin Shen, Wei Liu, Shaoting Zhang, Yang Yang, Heng Tao Shen
  • An Exploration of Parameter Redundancy in Deep Networks With Circulant Projections [pdf] - Yu Cheng, Felix X. Yu, Rogerio S. Feris, Sanjiv Kumar, Alok Choudhary, Shi-Fu Chang
  • Alternating Co-Quantization for Cross-Modal Hashing [pdf] - Go Irie, Hiroyuki Arai, Yukinobu Taniguchi
  • Projection Bank: From High-Dimensional Data to Medium-Length Binary Codes [pdf] - Li Liu, Mengyang Yu, Ling Shao
Group Membership Prediction [pdf]
Ziming Zhang, Yuting Chen, Venkatesh Saligrama

Abstract: The group membership prediction (GMP) problem involves predicting whether or not a collection of instances share a certain semantic property. For instance, in kinship verification given a collection of images, the goal is to predict whether or not they share a familial relationship. In this context we propose a novel probability model and introduce latent view-specific and view-shared random variables to jointly account for the view-specific appearance and crossview similarities among data instances. Our model posits that data from each view is independent conditioned on the shared variables. This postulate leads to a parametric probability model that decomposes group membership likelihood into a tensor product of data-independent parameters and data-dependent factors. We propose learning the data-independent parameters in a discriminative way with bilinear classifiers, and test our prediction algorithm on challenging visual recognition tasks such as multi-camera person re-identification and kinship verification. On most benchmark datasets, our method can significantly outperform the current state-of-the-art.
Similar papers:
  • A Spatio-Temporal Appearance Representation for Viceo-Based Pedestrian Re-Identification [pdf] - Kan Liu, Bingpeng Ma, Wei Zhang, Rui Huang
  • Partial Person Re-Identification [pdf] - Wei-Shi Zheng, Xiang Li, Tao Xiang, Shengcai Liao, Jianhuang Lai, Shaogang Gong
  • Multi-Scale Learning for Low-Resolution Person Re-Identification [pdf] - Xiang Li, Wei-Shi Zheng, Xiaojuan Wang, Tao Xiang, Shaogang Gong
  • Conditional High-Order Boltzmann Machine: A Supervised Learning Model for Relation Learning [pdf] - Yan Huang, Wei Wang, Liang Wang
  • Scalable Person Re-Identification: A Benchmark [pdf] - Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian
Higher-Order Inference for Multi-Class Log-Supermodular Models [pdf]
Jian Zhang, Josip Djolonga, Andreas Krause

Abstract: Higher-order models have been shown to be very useful for a plethora of computer vision tasks. However, existing techniques have focused mainly on MAP inference. In this paper, we present the first efficient approach towards approximate Bayesian marginal inference in a general class of high-order, multi-label attractive models, where previous techniques slow down exponentially with the order (clique size). We formalize this task as performing inference in log-supermodular models under partition constraints, and present an efficient variational inference technique. The resulting optimization problems are convex and yield bounds on the partition function. We also obtain a fully factorized approximation to the posterior, which can be used in lieu of the true complicated distribution. We empirically demonstrate the performance of our approach by comparing it to traditional inference methods on a challenging high-fidelity multi-label image segmentation dataset. We obtain stateof-the-art classification accuracy for MAP inference, and substantially improved ROC curves using the approximate marginals.
Similar papers:
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
  • Volumetric Bias in Segmentation and Reconstruction: Secrets and Solutions [pdf] - Yuri Boykov, Hossam Isack, Carl Olsson, Ismail Ben Ayed
  • Inferring M-Best Diverse Labelings in a Single One [pdf] - Alexander Kirillov, Bogdan Savchynskyy, Dmitrij Schlesinger, Dmitry Vetrov, Carsten Rother
  • Predicting Multiple Structured Visual Interpretations [pdf] - Debadeepta Dey, Varun Ramakrishna, Martial Hebert, J. Andrew Bagnell
  • A Multiscale Variable-Grouping Framework for MRF Energy Minimization [pdf] - Omer Meir, Meirav Galun, Stav Yagev, Ronen Basri, Irad Yavneh
Human Pose Estimation in Videos [pdf]
Dong Zhang, Mubarak Shah

Abstract: In this paper, we present a method to estimate a sequence of human poses in unconstrained videos. In contrast to the commonly employed graph optimization framework, which is NP-hard and needs approximate solutions, we formulate this problem into a unified two stage tree-based optimization problem for which an efficient and exact solution exists. Although the proposed method finds an exact solution, it does not sacrifice the ability to model the spatial and temporal constraints between body parts in the video frames; indeed it even models the symmetric parts better than the existing methods. The proposed method is based on two main ideas: Abstraction and Association to enforce the intra- and inter-frame body part constraints respectively without inducing extra computational complexity to the polynomial time solution. Using the idea of Abstraction, a new concept of abstract body part is introduced to model not only the tree based body part structure similar to existing methods, but also extra constraints between symmetric parts. Using the idea of Association, the optimal tracklets are generated for each abstract body part, in order to enforce the spatiotemporal constraints between body parts in adjacent frames. Finally, a sequence of the best poses is inferred from the abstract body part tracklets through the tree-based optimization. We evaluated the proposed method on three publicly available video based human pose estimation datasets, and obtained dramatically impro
Similar papers:
  • Detailed Full-Body Reconstructions of Moving People From Monocular RGB-D Sequences [pdf] - Federica Bogo, Michael J. Black, Matthew Loper, Javier Romero
  • Multi-Task Recurrent Neural Network for Immediacy Prediction [pdf] - Xiao Chu, Wanli Ouyang, Wei Yang, Xiaogang Wang
  • Beyond Tree Structure Models: A New Occlusion Aware Graphical Model for Human Pose Estimation [pdf] - Lianrui Fu, Junge Zhang, Kaiqi Huang
  • Recurrent Network Models for Human Dynamics [pdf] - Katerina Fragkiadaki, Sergey Levine, Panna Felsen, Jitendra Malik
  • Uncovering Interactions and Interactors: Joint Estimation of Head, Body Orientation and F-Formations From Surveillance Videos [pdf] - Elisa Ricci, Jagannadan Varadarajan, Ramanathan Subramanian, Samuel Rota Bulo, Narendra Ahuja, Oswald Lanz
Hyperspectral Compressive Sensing Using Manifold-Structured Sparsity Prior [pdf]
Lei Zhang, Wei Wei, Yanning Zhang, Fei Li, Chunhua Shen, Qinfeng Shi

Abstract: {chunhua.shen,javen.shi}@adelaide.edu.au 0.13 0.6 0.12 0.5 0.11 0.4 0.1 0.3 To reconstruct hyperspectral image (HSI) accurately from a few noisy compressive measurements, we present a novel manifold-structured sparsity prior based hyperspectral compressive sensing (HCS) method in this study. A matrix based hierarchical prior is first proposed to represent the spectral structured sparsity and spatial unknown manifold structure of HSI simultaneously. Then, a latent variable Bayes model is introduced to learn the sparsity prior and estimate the noise jointly from measurements. The learned prior can fully represent the inherent 3D structure of HSI and regulate its shape based on the estimated noise level. Thus, with this learned prior, the proposed method improves the reconstruction accuracy significantly and shows strong robustness to unknown noise in HCS. Experiments on four real hyperspectral datasets show that the proposed method outperforms several state-of-the-art methods on the reconstruction accuracy of HSI.
Similar papers:
  • Conformal and Low-Rank Sparse Representation for Image Restoration [pdf] - Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng
  • Low-Rank Matrix Factorization Under General Mixture Noise Distributions [pdf] - Xiangyong Cao, Yang Chen, Qian Zhao, Deyu Meng, Yao Wang, Dong Wang, Zongben Xu
  • RGB-Guided Hyperspectral Image Upsampling [pdf] - Hyeokhyen Kwon, Yu-Wing Tai
  • Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf] - Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • Hyperspectral Super-Resolution by Coupled Spectral Unmixing [pdf] - Charis Lanaras, Emmanuel Baltsavias, Konrad Schindler
Joint Camera Clustering and Surface Segmentation for Large-Scale Multi-View Stereo [pdf]
Runze Zhang, Shiwei Li, Tian Fang, Siyu Zhu, Long Quan

Abstract: In this paper, we propose an decomposition approach to large-scale multi-view stereo from an initial sparse reconstruction. The success of the approach depends on the introduction of surface-segmentation-based camera clustering rather than sparse-point-based camera clustering, which suffers from the problems of non-uniform reconstruction coverage ratio and high redundancy. In details, we introduce three criteria for camera clustering and surface segmentation for reconstruction, and then we formulate these criteria into an energy minimization problem under constraints. To solve this problem, we propose a joint optimization in a hierarchical framework to obtain the final surface segments and corresponding camera clusters. On each level of the hierarchical framework, the camera clustering problem is formulated as a parameter estimation problem of a probability model solved by a General Expectation-Maximization algorithm and the surface segmentation problem is formulated as a Markov Random Field model based on the probability estimated by the previous camera clustering process. The experiments on several Internet datasets and aerial photo datasets demonstrate that the proposed approach method generates more uniform and complete dense reconstruction with less redundancy, resulting in more efficient multi-view stereo algorithm.
Similar papers:
  • SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf] - Kensei Jo, Mohit Gupta, Shree K. Nayar
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
  • Higher-Order CRF Structural Segmentation of 3D Reconstructed Surfaces [pdf] - Jingbo Liu, Jinglu Wang, Tian Fang, Chiew-Lan Tai, Long Quan
  • General Dynamic Scene Reconstruction From Multiple View Video [pdf] - Armin Mustafa, Hansung Kim, Jean-Yves Guillemaut, Adrian Hilton
Learning Social Relation Traits From Face Images [pdf]
Zhanpeng Zhang, Ping Luo, Chen-Change Loy, Xiaoou Tang

Abstract: Social relation denes the association, e.g., warm, friendliness, and dominance, between two or more people. Motivated by psychological studies, we investigate if such ne-grained and high-level relation traits can be characterised and quantied from face images in the wild. To address this challenging problem we propose a deep model that learns a rich face representation to capture gender, expression, head pose, and age-related attributes, and then performs pairwise-face reasoning for relation prediction. To learn from heterogeneous attribute sources, we formulate a new network architecture with a bridging layer to leverage the inherent correspondences among these datasets. It can also cope with missing target attribute labels. Extensive experiments show that our approach is effective for ne-grained social relation learning in images and videos.
Similar papers:
  • From Facial Parts Responses to Face Detection: A Deep Learning Approach [pdf] - Shuo Yang, Ping Luo, Chen-Change Loy, Xiaoou Tang
  • Learning Image and User Features for Recommendation in Social Networks [pdf] - Xue Geng, Hanwang Zhang, Jingwen Bian, Tat-Seng Chua
  • Deep Learning Face Attributes in the Wild [pdf] - Ziwei Liu, Ping Luo, Xiaogang Wang, Xiaoou Tang
  • Conditional High-Order Boltzmann Machine: A Supervised Learning Model for Relation Learning [pdf] - Yan Huang, Wei Wang, Liang Wang
  • Automated Facial Trait Judgment and Election Outcome Prediction: Social Dimensions of Face [pdf] - Jungseock Joo, Francis F. Steen, Song-Chun Zhu
Leveraging Datasets With Varying Annotations for Face Alignment via Deep Regression Network [pdf]
Jie Zhang, Meina Kan, Shiguang Shan, Xilin Chen

Abstract: Facial landmark detection, as a vital topic in computer vision, has been studied for many decades and lots of datasets have been collected for evaluation. These datasets usually have different annotations, e.g., 68-landmark markup for LFPW dataset, while 74-landmark markup for GTAV dataset. Intuitively, it is meaningful to fuse all the datasets to predict a union of all types of landmarks from multiple datasets (i.e., transfer the annotations of each dataset to all other datasets), but this problem is nontrivial due to the distribution discrepancy between datasets and incomplete annotations of all types for each dataset. In this work, we propose a deep regression network coupled with sparse shape regression (DRN-SSR) to predict the union of all types of landmarks by leveraging datasets with varying annotations, each dataset with one type of annotation. Specifically, the deep regression network intends to predict the union of all landmarks, and the sparse shape regression attempts to approximate those undefined landmarks on each dataset so as to guide the learning of the deep regression network for face alignment. Extensive experiments on two challenging datasets, IBUG and GLF, demonstrate that our method can effectively leverage multiple datasets with different annotations to predict the union of all landmarks.
Similar papers:
  • Single Image Pop-Up From Discriminatively Learned Parts [pdf] - Menglong Zhu, Xiaowei Zhou, Kostas Daniilidis
  • The One Triangle Three Parallelograms Sampling Strategy and Its Application in Shape Regression [pdf] - Mikael Nilsson
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
  • Pose-Invariant 3D Face Alignment [pdf] - Amin Jourabloo, Xiaoming Liu
  • Regressive Tree Structured Model for Facial Landmark Localization [pdf] - Gee-Sern Hsu, Kai-Hsiang Chang, Shih-Chieh Huang
Low-Rank Tensor Constrained Multiview Subspace Clustering [pdf]
Changqing Zhang, Huazhu Fu, Si Liu, Guangcan Liu, Xiaochun Cao

Abstract: In this paper, we explore the problem of multiview subspace clustering. We introduce a low-rank tensor constraint to explore the complementary information from multiple views and, accordingly, establish a novel method called Low-rank Tensor constrained Multiview Subspace Clustering (LT-MSC). Our method regards the subspace representation matrices of different views as a tensor, which captures dexterously the high order correlations underlying multiview data. Then the tensor is equipped with a low-rank constraint, which models elegantly the cross information among different views, reduces effectually the redundancy of the learned subspace representations, and improves the accuracy of clustering as well. The inference process of the affinity matrix for clustering is formulated as a tensor nuclear norm minimization problem, constrained with an additional 2,1 -norm regularizer and some linear equalities. The minimization problem is convex and thus can be solved efficiently by an Augmented Lagrangian Alternating Direction Minimization (AL-ADM) method. Extensive experimental results on four benchmark image datasets show the effectiveness of the proposed LT-MSC method.
Similar papers:
  • Low-Rank Tensor Approximation With Laplacian Scale Mixture Modeling for Multiframe Image Denoising [pdf] - Weisheng Dong, Guangyu Li, Guangming Shi, Xin Li, Yi Ma
  • Shape Interaction Matrix Revisited and Robustified: Efficient Subspace Clustering With Corrupted and Incomplete Data [pdf] - Pan Ji, Mathieu Salzmann, Hongdong Li
  • A Novel Sparsity Measure for Tensor Recovery [pdf] - Qian Zhao, Deyu Meng, Xu Kong, Qi Xie, Wenfei Cao, Yao Wang, Zongben Xu
  • Temporal Subspace Clustering for Human Motion Segmentation [pdf] - Sheng Li, Kang Li, Yun Fu
  • Multi-View Subspace Clustering [pdf] - Hongchang Gao, Feiping Nie, Xuelong Li, Heng Huang
MeshStereo: A Global Stereo Model With Mesh Alignment Regularization for View Interpolation [pdf]
Chi Zhang, Zhiwei Li, Yanhua Cheng, Rui Cai, Hongyang Chao, Yong Rui

Abstract: We present a novel global stereo model designed for view interpolation. Unlike existing stereo models which only output a disparity map, our model is able to output a 3D triangular mesh, which can be directly used for view interpolation. To this aim, we partition the input stereo images into 2D triangles with shared vertices. Lifting the 2D triangulation to 3D naturally generates a corresponding mesh. A technical difficulty is to properly split vertices to multiple copies when they appear at depth discontinuous boundaries. To deal with this problem, we formulate our objective as a two-layer MRF, with the upper layer modeling the splitting properties of the vertices and the lower layer optimizing a region-based stereo matching. Experiments on the Middlebury and the Herodion datasets demonstrate that our model is able to synthesize visually coherent new view angles with high PSNR, as well as outputting high quality disparity maps which rank at the first place on the new challenging high resolution Middlebury 3.0 benchmark.
Similar papers:
  • Depth Map Estimation and Colorization of Anaglyph Images Using Local Color Prior and Reverse Intensity Distribution [pdf] - Williem, Ramesh Raskar, In Kyu Park
  • Blur-Aware Disparity Estimation From Defocus Stereo Images [pdf] - Ching-Hui Chen, Hui Zhou, Timo Ahonen
  • A Deep Visual Correspondence Embedding Model for Stereo Matching Costs [pdf] - Zhuoyuan Chen, Xun Sun, Liang Wang, Yinan Yu, Chang Huang
  • The HCI Stereo Metrics: Geometry-Aware Performance Analysis of Stereo Algorithms [pdf] - Katrin Honauer, Lena Maier-Hein, Daniel Kondermann
  • MAP Disparity Estimation Using Hidden Markov Trees [pdf] - Eric T. Psota, Jedrzej Kowalczuk, Mateusz Mittek, Lance C. Perez
Minimum Barrier Salient Object Detection at 80 FPS [pdf]
Jianming Zhang, Stan Sclaroff, Zhe Lin, Xiaohui Shen, Brian Price, Radomir Mech

Abstract: We propose a highly efficient, yet powerful, salient object detection method based on the Minimum Barrier Distance (MBD) Transform. The MBD transform is robust to pixelvalue fluctuation, and thus can be effectively applied on raw pixels without region abstraction. We present an approximate MBD transform algorithm with 100X speedup over the exact algorithm. An error bound analysis is also provided. Powered by this fast MBD transform algorithm, the proposed salient object detection method runs at 80 FPS, and significantly outperforms previous methods with similar speed on four large benchmark datasets, and achieves comparable or better performance than state-of-the-art methods. Furthermore, a technique based on color whitening is proposed to extend our method to leverage the appearancebased backgroundness cue. This extended version further improves the performance, while still being one order of magnitude faster than all the other leading methods.
Similar papers:
  • A Data-Driven Metric for Comprehensive Evaluation of Saliency Models [pdf] - Jia Li, Changqun Xia, Yafei Song, Shu Fang, Xiaowu Chen
  • Generic Promotion of Diffusion-Based Salient Object Detection [pdf] - Peng Jiang, Nuno Vasconcelos, Jingliang Peng
  • SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf] - Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao
  • Cluster-Based Point Set Saliency [pdf] - Flora Ponjou Tasse, Jiri Kosinka, Neil Dodgson
  • HARF: Hierarchy-Associated Rich Features for Salient Object Detection [pdf] - Wenbin Zou, Nikos Komodakis
Mining And-Or Graphs for Graph Matching and Object Discovery [pdf]
Quanshi Zhang, Ying Nian Wu, Song-Chun Zhu

Abstract: This paper reformulates the theory of graph mining on the technical basis of graph matching, and extends its scope of applications to computer vision. Given a set of attributed relational graphs (ARGs), we propose to use a hierarchical And-Or Graph (AoG) to model the pattern of maximal-size common subgraphs embedded in the ARGs, and we develop a general method to mine the AoG model from the unlabeled ARGs. This method provides a general solution to the problem of mining hierarchical models from unannotated visual data without exhaustive search of objects. We apply our method to RGB/RGB-D images and videos to demonstrate its generality and the wide range of applicability. The code will be available at https://sites.google.com/site/ quanshizhang/mining-and-or-graphs.
Similar papers:
  • Probabilistic Label Relation Graphs With Ising Models [pdf] - Nan Ding, Jia Deng, Kevin P. Murphy, Hartmut Neven
  • Weakly Supervised Graph Based Semantic Segmentation by Learning Communities of Image-Parts [pdf] - Niloufar Pourian, S. Karthikeyan, B.S. Manjunath
  • Attribute-Graph: A Graph Based Approach to Image Ranking [pdf] - Nikita Prabhu, R. Venkatesh Babu
  • Discrete Tabu Search for Graph Matching [pdf] - Kamil Adamczewski, Yumin Suh, Kyoung Mu Lee
  • A Matrix Decomposition Perspective to Multiple Graph Matching [pdf] - Junchi Yan, Hongteng Xu, Hongyuan Zha, Xiaokang Yang, Huanxi Liu, Stephen Chu
Monocular Object Instance Segmentation and Depth Ordering With CNNs [pdf]
Ziyu Zhang, Alexander G. Schwing, Sanja Fidler, Raquel Urtasun

Abstract: In this paper we tackle the problem of instance-level segmentation and depth ordering from a single monocular image. Towards this goal, we take advantage of convolutional neural nets and train them to directly predict instance-level segmentations where the instance ID encodes the depth ordering within image patches. To provide a coherent single explanation of an image we develop a Markov random field which takes as input the predictions of convolutional neural nets applied at overlapping patches of different resolutions, as well as the output of a connected component algorithm. It aims to predict accurate instance-level segmentation and depth ordering. We demonstrate the effectiveness of our approach on the challenging KITTI benchmark and show good performance on both tasks.
Similar papers:
  • Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf] - Naejin Kong, Michael J. Black
  • AttentionNet: Aggregating Weak Directions for Accurate Object Detection [pdf] - Donggeun Yoo, Sunggyun Park, Joon-Young Lee, Anthony S. Paek, In So Kweon
  • Predicting Depth, Surface Normals and Semantic Labels With a Common Multi-Scale Convolutional Architecture [pdf] - David Eigen, Rob Fergus
  • Relaxed Multiple-Instance SVM With Application to Object Discovery [pdf] - Xinggang Wang, Zhuotun Zhu, Cong Yao, Xiang Bai
  • Simpler Non-Parametric Methods Provide as Good or Better Results to Multiple-Instance Learning [pdf] - Ragav Venkatesan, Parag Chandakkar, Baoxin Li
Segment Graph Based Image Filtering: Fast Structure-Preserving Smoothing [pdf]
Feihu Zhang, Longquan Dai, Shiming Xiang, Xiaopeng Zhang

Abstract: In this paper, we design a new edge-aware structure, named segment graph, to represent the image and we further develop a novel double weighted average image filter (SGF) based on the segment graph. In our SGF, we use the tree distance on the segment graph to define the internal weight function of the filtering kernel, which enables the filter to smooth out high-contrast details and textures while preserving major image structures very well. While for the external weight function, we introduce a user specified smoothing window to balance the smoothing effects from each node of the segment graph. Moreover, we also set a threshold to adjust the edge-preserving performance. These advantages make the SGF more flexible in various applications and overcome the halo and leak problems appearing in most of the state-of-the-art approaches. Finally and importantly, we develop a linear algorithm for the implementation of our SGF, which has an O(N ) time complexity for both gray-scale and high dimensional images, regardless of the kernel size and the intensity range. Typically, as one of the fastest edge-preserving filters, our CPU implementation achieves 0.15s per megapixel when performing filtering for 3-channel color images. The strength of the proposed filter is demonstrated by various applications, including stereo matching, optical flow, joint depth map upsampling, edge-preserving smoothing, edges detection, image denoising, abstraction and texture editing.
Similar papers:
  • Efficient Video Segmentation Using Parametric Graph Partitioning [pdf] - Chen-Ping Yu, Hieu Le, Gregory Zelinsky, Dimitris Samaras
  • SPM-BP: Sped-up PatchMatch Belief Propagation for Continuous MRFs [pdf] - Yu Li, Dongbo Min, Michael S. Brown, Minh N. Do, Jiangbo Lu
  • Adaptive Exponential Smoothing for Online Filtering of Pixel Prediction Maps [pdf] - Kang Dang, Jiong Yang, Junsong Yuan
  • Mutual-Structure for Joint Filtering [pdf] - Xiaoyong Shen, Chao Zhou, Li Xu, Jiaya Jia
  • Fully Connected Guided Image Filtering [pdf] - Longquan Dai, Mengke Yuan, Feihu Zhang, Xiaopeng Zhang
Shell PCA: Statistical Shape Modelling in Shell Space [pdf]
Chao Zhang, Behrend Heeren, Martin Rumpf, William A. P. Smith

Abstract: are plausibly members of the object class being modelled. In this paper we describe how to perform Principal Components Analysis in shell space. Thin shells are a physical model for surfaces with non-zero thickness whose deformation dissipates elastic energy. Thin shells, or their discrete counterparts, can be considered to reside in a shell space in which the notion of distance is given by the elastic energy required to deform one shape into another. It is in this setting that we show how to perform statistical analysis of a set of shapes (meshes in dense correspondence), providing a hybrid between physical and statistical shape modelling. The resulting models are better able to capture nonlinear deformations, for example resulting from articulated motion, even when training data is very sparse compared to the dimensionality of the observation space. Often, the dimensionality of the observations is orders of magnitude greater than the number of samples in the training set. With modern shape capture techniques, typical meshes may contain tens of thousands of vertices. In contrast, the number of training samples that can feasibly be collected is typically only in the tens or hundreds. Hence, the underlying shape space is sampled very sparsely.
Similar papers:
  • BodyPrint: Pose Invariant 3D Shape Matching of Human Bodies [pdf] - Jiangping Wang, Kai Ma, Vivek Kumar Singh, Thomas Huang, Terrence Chen
  • A Groupwise Multilinear Correspondence Optimization for 3D Faces [pdf] - Timo Bolkart, Stefanie Wuhrer
  • Detailed Full-Body Reconstructions of Moving People From Monocular RGB-D Sequences [pdf] - Federica Bogo, Michael J. Black, Matthew Loper, Javier Romero
  • Compositional Hierarchical Representation of Shape Manifolds for Classification of Non-Manifold Shapes [pdf] - Mete Ozay, Umit Rusen Aktas, Jeremy L. Wyatt, Ales Leonardis
  • Learning Shape, Motion and Elastic Models in Force Space [pdf] - Antonio Agudo, Francesc Moreno-Noguer
Zero-Shot Learning via Semantic Similarity Embedding [pdf]
Ziming Zhang, Venkatesh Saligrama

Abstract: DOG TREE BOAT TRUCK CAT Histogram of Unseen Source Mixtures BOAT TREE TRUCK CAT CAR TRUCK HORSE BOAT CAR BOAT TREE TRUCK CAT HORSE While there has been significant progress in large-scale classification in recent years [31], lack of sufficient training data for every class and the increasing difficulty in finding annotations for a large fraction of data might impact further improvements. Zero-shot learning is being increasingly recognized as a way to deal with these difficulties. One version of zero shot learning is based on so-called source and target domains. Source domain is described by a single vector corresponding to each class based on side information such as attributes [8, 16, 21, 25, 29], language words/phrases [4, 9, 34], or even learned classifiers [42], which we assume can be collected easily. The target domain is described by a joint distribution of images/videos and labels [16, 41]. During training time, we are given source domain attributes and target domain data corresponding to only a subset of classes, which we call seen classes. During test time, source domain attributes for unseen (i.e. no training data provided) CAT
Similar papers:
  • Multi-View Domain Generalization for Visual Recognition [pdf] - Li Niu, Wen Li, Dong Xu
  • Unsupervised Domain Adaptation With Imbalanced Cross-Domain Data [pdf] - Tzu Ming Harry Hsu, Wei Yu Chen, Cheng-An Hou, Yao-Hung Hubert Tsai, Yi-Ren Yeh, Yu-Chiang Frank Wang
  • Simultaneous Deep Transfer Across Domains and Tasks [pdf] - Eric Tzeng, Judy Hoffman, Trevor Darrell, Kate Saenko
  • Bi-Shifting Auto-Encoder for Unsupervised Domain Adaptation [pdf] - Meina Kan, Shiguang Shan, Xilin Chen
  • Unsupervised Domain Adaptation for Zero-Shot Learning [pdf] - Elyor Kodirov, Tao Xiang, Zhenyong Fu, Shaogang Gong
A Novel Sparsity Measure for Tensor Recovery [pdf]
Qian Zhao, Deyu Meng, Xu Kong, Qi Xie, Wenfei Cao, Yao Wang, Zongben Xu

Abstract: In this paper, we propose a new sparsity regularizer for measuring the low-rank structure underneath a tensor. The proposed sparsity measure has a natural physical meaning which is intrinsically the size of the fundamental Kronecker basis to express the tensor. By embedding the sparsity measure into the tensor completion and tensor robust PCA frameworks, we formulate new models to enhance their capability in tensor recovery. Through introducing relaxation forms of the proposed sparsity measure, we also adopt the alternating direction method of multipliers (ADMM) for solving the proposed models. Experiments implemented on synthetic and multispectral image data sets substantiate the effectiveness of the proposed methods.
Similar papers:
  • A Groupwise Multilinear Correspondence Optimization for 3D Faces [pdf] - Timo Bolkart, Stefanie Wuhrer
  • Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf] - Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf] - Yuhui Quan, Yan Huang, Hui Ji
  • Low-Rank Tensor Constrained Multiview Subspace Clustering [pdf] - Changqing Zhang, Huazhu Fu, Si Liu, Guangcan Liu, Xiaochun Cao
  • Low-Rank Tensor Approximation With Laplacian Scale Mixture Modeling for Multiframe Image Denoising [pdf] - Weisheng Dong, Guangyu Li, Guangming Shi, Xin Li, Yi Ma
An Accurate Iris Segmentation Framework Under Relaxed Imaging Constraints Using Total Variation Model [pdf]
Zijing Zhao, Kumar Ajay

Abstract: This paper proposes a novel and more accurate iris segmentation framework to automatically segment iris region from the face images acquired with relaxed imaging under visible or near-infrared illumination, which provides strong feasibility for applications in surveillance, forensics and the search for missing children, etc. The proposed framework is built on a novel total-variation based formulation which uses l1 norm regularization to robustly suppress noisy texture pixels for the accurate iris localization. A series of novel and robust post processing operations are introduced to more accurately localize the limbic boundaries. Our experimental results on three publicly available databases, i.e., FRGC, UBIRIS.v2 and CASIA.v4-distance, achieve significant performance improvement in terms of iris segmentation accuracy over the state-of-the-art approaches in the literature. Besides, we have shown that using iris masks generated from the proposed approach helps to improve iris recognition performance as well. Unlike prior work, all the implementations in this paper are made publicly available to further advance research and applications in biometrics at-d-distance.
Similar papers:
  • Object Detection via a Multi-Region and Semantic Segmentation-Aware CNN Model [pdf] - Spyros Gidaris, Nikos Komodakis
  • BoxSup: Exploiting Bounding Boxes to Supervise Convolutional Networks for Semantic Segmentation [pdf] - Jifeng Dai, Kaiming He, Jian Sun
  • Efficient Classifier Training to Minimize False Merges in Electron Microscopy Segmentation [pdf] - Toufiq Parag, Dan C. Ciresan, Alessandro Giusti
  • A Randomized Ensemble Approach to Industrial CT Segmentation [pdf] - Hyojin Kim, Jayaraman Jayaraman J. Thiagarajan, Peer-Timo Bremer
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
Conditional Random Fields as Recurrent Neural Networks [pdf]
Shuai Zheng, Sadeep Jayasumana, Bernardino Romera-Paredes, Vibhav Vineet, Zhizhong Su, Dalong Du, Chang Huang, Philip H. S. Torr

Abstract: Pixel-level labelling tasks, such as semantic segmentation, play a central role in image understanding. Recent approaches have attempted to harness the capabilities of deep learning techniques for image recognition to tackle pixellevel labelling tasks. One central issue in this methodology is the limited capacity of deep learning techniques to delineate visual objects. To solve this problem, we introduce a new form of convolutional neural network that combines the strengths of Convolutional Neural Networks (CNNs) and Conditional Random Fields (CRFs)-based probabilistic graphical modelling. To this end, we formulate Conditional Random Fields with Gaussian pairwise potentials and mean-field approximate inference as Recurrent Neural Networks. This network, called CRF-RNN, is then plugged in as a part of a CNN to obtain a deep network that has desirable properties of both CNNs and CRFs. Importantly, our system fully integrates CRF modelling with CNNs, making it possible to train the whole deep network end-to-end with the usual back-propagation algorithm, avoiding offline post-processing methods for object delineation. We apply the proposed method to the problem of semantic image segmentation, obtaining top results on the challenging Pascal VOC 2012 segmentation benchmark.
Similar papers:
  • Hierarchical Higher-Order Regression Forest Fields: An Application to 3D Indoor Scene Labelling [pdf] - Trung T. Pham, Ian Reid, Yasir Latif, Stephen Gould
  • Semantic Segmentation With Object Clique Potential [pdf] - Xiaojuan Qi, Jianping Shi, Shu Liu, Renjie Liao, Jiaya Jia
  • Joint Object and Part Segmentation Using Deep Learned Potentials [pdf] - Peng Wang, Xiaohui Shen, Zhe Lin, Scott Cohen, Brian Price, Alan L. Yuille
  • Learning Deconvolution Network for Semantic Segmentation [pdf] - Hyeonwoo Noh, Seunghoon Hong, Bohyung Han
  • Semantic Image Segmentation via Deep Parsing Network [pdf] - Ziwei Liu, Xiaoxiao Li, Ping Luo, Chen-Change Loy, Xiaoou Tang
Minimal Solvers for 3D Geometry From Satellite Imagery [pdf]
Enliang Zheng, Ke Wang, Enrique Dunn, Jan-Michael Frahm

Abstract: (X, Y, Z) (X, Y, Z) Mapping (X, Y, Z) We propose two novel minimal solvers which advance Inverse Mapping Inverse using RFM the state of the art in satellite imagery processing. Our mapping using RPC mapping methods are efficient and do not rely on the prior existence of complex inverse mapping functions to correlate 2D im(r, c) (r2,c(r21),c1) (r1,c1) (r, improves c) age coordinates and 3D terrain. Our first solver on the stereo correspondence problem for satellite imagery, (a) Inverse mapping (b) Triangulation in that we provide an exact image-to-object space mapping (where prior methods were inaccurate). Our second Figure 1: The two problems solved in our paper. solver provides a novel mechanism for 3D point triangulation, which has improved robustness and accuracy over prior techniques. Given the usefulness and ubiquity of satelincluded in the standards of the NITF image format [1], and lite imagery, our proposed methods allow for improved reis available from most satellite image vendors [2]. sults in a variety of existing and future applications. However, one challenge faced by many works utilizing satellite imagery and the RPC model is that the bidirectional mapping between the 3D object and 2D image spaces
Similar papers:
  • Exploiting High Level Scene Cues in Stereo Reconstruction [pdf] - Simon Hadfield, Richard Bowden
  • Wide Baseline Stereo Matching With Convex Bounded Distortion Constraints [pdf] - Meirav Galun, Tal Amir, Tal Hassner, Ronen Basri, Yaron Lipman
  • Efficient Solution to the Epipolar Geometry for Radially Distorted Cameras [pdf] - Zuzana Kukelova, Jan Heller, Martin Bujnak, Andrew Fitzgibbon, Tomas Pajdla
  • An Efficient Minimal Solution for Multi-Camera Motion [pdf] - Jonathan Ventura, Clemens Arth, Vincent Lepetit
  • Structure From Motion Using Structure-Less Resection [pdf] - Enliang Zheng, Changchang Wu
Partial Person Re-Identification [pdf]
Wei-Shi Zheng, Xiang Li, Tao Xiang, Shengcai Liao, Jianhuang Lai, Shaogang Gong

Abstract: We address a new partial person re-identication (reid) problem, where only a partial observation of a person is available for matching across different non-overlapping camera views. This differs significantly from the conventional person re-id setting where it is assumed that the full body of a person is detected and aligned. To solve this more challenging and realistic re-id problem without the implicit assumption of manual body-parts alignment, we propose a matching framework consisting of 1) a local patch-level matching model based on a novel sparse representation classification formulation with explicit patch ambiguity modelling, and 2) a global part-based matching model providing complementary spatial layout information. Our framework is evaluated on a new partial person re-id dataset as well as two existing datasets modified to include partial person images. The results show that the proposed method outperforms significantly existing re-id methods as well as other partial visual matching methods.
Similar papers:
  • Scalable Person Re-Identification: A Benchmark [pdf] - Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian
  • Person Re-Identification Ranking Optimisation by Discriminant Context Information Analysis [pdf] - Jorge Garcia, Niki Martinel, Christian Micheloni, Alfredo Gardel
  • Person Re-Identification With Correspondence Structure Learning [pdf] - Yang Shen, Weiyao Lin, Junchi Yan, Mingliang Xu, Jianxin Wu, Jingdong Wang
  • Person Re-Identification With Discriminatively Trained Viewpoint Invariant Dictionaries [pdf] - Srikrishna Karanam, Yang Li, Richard J. Radke
  • Multi-Scale Learning for Low-Resolution Person Re-Identification [pdf] - Xiang Li, Wei-Shi Zheng, Xiaojuan Wang, Tao Xiang, Shaogang Gong
Scalable Person Re-Identification: A Benchmark [pdf]
Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, Qi Tian

Abstract: This paper contributes a new high quality dataset for person re-identification, named Market-1501. Generally, current datasets: 1) are limited in scale; 2) consist of hand-drawn bboxes, which are unavailable under realistic settings; 3) have only one ground truth and one query image for each identity (close environment). To tackle these problems, the proposed Market-1501 dataset is featured in three aspects. First, it contains over 32,000 annotated bboxes, plus a distractor set of over 500K images, making it the largest person re-id dataset to date. Second, images in Market-1501 dataset are produced using the Deformable Part Model (DPM) as pedestrian detector. Third, our dataset is collected in an open system, where each identity has multiple images under each camera. As a minor contribution, inspired by recent advances in large-scale image search, this paper proposes an unsupervised Bag-of-Words descriptor. We view person reidentification as a special task of image search. In experiment, we show that the proposed descriptor yields competitive accuracy on VIPeR, CUHK03, and Market-1501 datasets, and is scalable on the large-scale 500k dataset.
Similar papers:
  • A Spatio-Temporal Appearance Representation for Viceo-Based Pedestrian Re-Identification [pdf] - Kan Liu, Bingpeng Ma, Wei Zhang, Rui Huang
  • Partial Person Re-Identification [pdf] - Wei-Shi Zheng, Xiang Li, Tao Xiang, Shengcai Liao, Jianhuang Lai, Shaogang Gong
  • Learning Query and Image Similarities With Ranking Canonical Correlation Analysis [pdf] - Ting Yao, Tao Mei, Chong-Wah Ngo
  • Multi-Scale Learning for Low-Resolution Person Re-Identification [pdf] - Xiang Li, Wei-Shi Zheng, Xiaojuan Wang, Tao Xiang, Shaogang Gong
  • Efficient PSD Constrained Asymmetric Metric Learning for Person Re-Identification [pdf] - Shengcai Liao, Stan Z. Li
Separating Fluorescent and Reflective Components by Using a Single Hyperspectral Image [pdf]
Yinqiang Zheng, Ying Fu, Antony Lam, Imari Sato, Yoichi Sato

Abstract: This paper introduces a novel method to separate fluorescent and reflective components in the spectral domain. In contrast to existing methods, which require to capture two or more images under varying illuminations, we aim to achieve this separation task by using a single hyperspectral image. After identifying the critical hurdle in singleimage component separation, we mathematically design the optimal illumination spectrum, which is shown to contain substantial high-frequency components in the frequency domain. This observation, in turn, leads us to recognize a key difference between reflectance and fluorescence in response to the frequency modulation effect of illumination, which fundamentally explains the feasibility of our method. On the practical side, we successfully find an off-the-shelf lamp as the light source, which is strong in irradiance intensity and cheap in cost. A fast linear separation algorithm is developed as well. Experiments using both synthetic data and real images have confirmed the validity of the selected illuminant and the accuracy of our separation algorithm.
Similar papers:
  • Adaptive Spatial-Spectral Dictionary Learning for Hyperspectral Image Denoising [pdf] - Ying Fu, Antony Lam, Imari Sato, Yoichi Sato
  • A Gaussian Process Latent Variable Model for BRDF Inference [pdf] - Stamatios Georgoulis, Vincent Vanweddingen, Marc Proesmans, Luc Van Gool
  • Learning Data-Driven Reflectance Priors for Intrinsic Image Decomposition [pdf] - Tinghui Zhou, Philipp Krahenbuhl, Alexei A. Efros
  • Hyperspectral Super-Resolution by Coupled Spectral Unmixing [pdf] - Charis Lanaras, Emmanuel Baltsavias, Konrad Schindler
  • RGB-Guided Hyperspectral Image Upsampling [pdf] - Hyeokhyen Kwon, Yu-Wing Tai
Sparse Dynamic 3D Reconstruction From Unsynchronized Videos [pdf]
Enliang Zheng, Dinghuang Ji, Enrique Dunn, Jan-Michael Frahm

Abstract: We target the sparse 3D reconstruction of dynamic objects observed by multiple unsynchronized video cameras with unknown temporal overlap. To this end, we develop a framework to recover the unknown structure without sequencing information across video sequences. Our proposed compressed sensing framework poses the estimation of 3D structure as the problem of dictionary learning. Moreover, we define our dictionary as the temporally varying 3D structure, while we define local sequencing information in terms of the sparse coefficients describing a locally linear 3D structural interpolation. Our formulation optimizes a biconvex cost function that leverages a compressed sensing formulation and enforces both structural dependency coherence across video streams, as well as motion smoothness across estimates from common video sources. Experimental results demonstrate the effectiveness of our approach in both synthetic data and captured imagery.
Similar papers:
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
  • General Dynamic Scene Reconstruction From Multiple View Video [pdf] - Armin Mustafa, Hansung Kim, Jean-Yves Guillemaut, Adrian Hilton
  • Integrating Dashcam Views Through Inter-Video Mapping [pdf] - Hsin-I Chen, Yi-Ling Chen, Wei-Tse Lee, Fan Wang, Bing-Yu Chen
  • Learning Shape, Motion and Elastic Models in Force Space [pdf] - Antonio Agudo, Francesc Moreno-Noguer
Structure From Motion Using Structure-Less Resection [pdf]
Enliang Zheng, Changchang Wu

Abstract: This paper proposes a new incremental structure from motion (SfM) algorithm based on a novel structure-less camera resection technique. Traditional methods rely on 2D-3D correspondences to compute the pose of candidate cameras using PnP. In this work, we take the collection of already reconstructed cameras as a generalized camera, and determine the absolute pose of a candidate pinhole camera from pure 2D correspondences, which we call it semi-generalized camera pose problem. We present the minimal solvers of the new problem for both calibrated and partially calibrated (unknown focal length) pinhole cameras. By integrating these new algorithms in an incremental SfM system, we go beyond the state-of-art methods with the capability of reconstructing cameras without 2D-3D correspondences. Large-scale real image experiments show that our new SfM system significantly improves the completeness of 3D reconstruction over the standard approach. O2 O1 O2 O1 O3 O3
Similar papers:
  • Rolling Shutter Super-Resolution [pdf] - Abhijith Punnappurath, Vijay Rengarajan, A.N. Rajagopalan
  • SpeDo: 6 DOF Ego-Motion Sensor Using Speckle Defocus Imaging [pdf] - Kensei Jo, Mohit Gupta, Shree K. Nayar
  • Minimal Solvers for 3D Geometry From Satellite Imagery [pdf] - Enliang Zheng, Ke Wang, Enrique Dunn, Jan-Michael Frahm
  • Efficient Solution to the Epipolar Geometry for Radially Distorted Cameras [pdf] - Zuzana Kukelova, Jan Heller, Martin Bujnak, Andrew Fitzgibbon, Tomas Pajdla
  • An Efficient Minimal Solution for Multi-Camera Motion [pdf] - Jonathan Ventura, Clemens Arth, Vincent Lepetit
Exploiting Object Similarity in 3D Reconstruction [pdf]
Chen Zhou, Fatma Guney, Yizhou Wang, Andreas Geiger

Abstract: Despite recent progress, reconstructing outdoor scenes in 3D from movable platforms remains a highly difficult endeavour. Challenges include low frame rates, occlusions, large distortions and difficult lighting conditions. In this paper, we leverage the fact that the larger the reconstructed area, the more likely objects of similar type and shape will occur in the scene. This is particularly true for outdoor scenes where buildings and vehicles often suffer from missing texture or reflections, but share similarity in 3D shape. We take advantage of this shape similarity by localizing objects using detectors and jointly reconstructing them while learning a volumetric model of their shape. This allows us to reduce noise while completing missing surfaces as objects of similar shape benefit from all observations for the respective category. We evaluate our approach with respect to LIDAR ground truth on a novel challenging suburban dataset and show its advantages over the state-of-the-art.
Similar papers:
  • Direct, Dense, and Deformable: Template-Based Non-Rigid 3D Reconstruction From RGB Video [pdf] - Rui Yu, Chris Russell, Neill D. F. Campbell, Lourdes Agapito
  • As-Rigid-As-Possible Volumetric Shape-From-Template [pdf] - Shaifali Parashar, Daniel Pizarro, Adrien Bartoli, Toby Collins
  • Deformable 3D Fusion: From Partial Dynamic 3D Observations to Complete 4D Models [pdf] - Weipeng Xu, Mathieu Salzmann, Yongtian Wang, Yue Liu
  • Global, Dense Multiscale Reconstruction for a Billion Points [pdf] - Benjamin Ummenhofer, Thomas Brox
  • General Dynamic Scene Reconstruction From Multiple View Video [pdf] - Armin Mustafa, Hansung Kim, Jean-Yves Guillemaut, Adrian Hilton
Learning Data-Driven Reflectance Priors for Intrinsic Image Decomposition [pdf]
Tinghui Zhou, Philipp Krahenbuhl, Alexei A. Efros

Abstract: We propose a data-driven approach for intrinsic image decomposition, which is the process of inferring the confounding factors of reflectance and shading in an image. We pose this as a two-stage learning problem. First, we train a model to predict relative reflectance ordering between image patches (brighter, darker, same) from large-scale human annotations, producing a data-driven reflectance prior. Second, we show how to naturally integrate this learned prior into existing energy minimization frameworks for intrinsic image decomposition. We compare our method to the state-of-the-art approach of Bell et al. [7] on both decomposition and image relighting tasks, demonstrating the benefits of the simple relative reflectance prior, especially for scenes under challenging lighting conditions. X Y (a) Original image
Similar papers:
  • Intrinsic Scene Decomposition From RGB-D images [pdf] - Mohammed Hachama, Bernard Ghanem, Peter Wonka
  • A Comprehensive Multi-Illuminant Dataset for Benchmarking of the Intrinsic Image Algorithms [pdf] - Shida Beigpour, Andreas Kolb, Sven Kunz
  • A Gaussian Process Latent Variable Model for BRDF Inference [pdf] - Stamatios Georgoulis, Vincent Vanweddingen, Marc Proesmans, Luc Van Gool
  • Learning Ordinal Relationships for Mid-Level Vision [pdf] - Daniel Zoran, Phillip Isola, Dilip Krishnan, William T. Freeman
  • Intrinsic Decomposition of Image Sequences From Local Temporal Variations [pdf] - Pierre-Yves Laffont, Jean-Charles Bazin
Multi-Image Matching via Fast Alternating Minimization [pdf]
Xiaowei Zhou, Menglong Zhu, Kostas Daniilidis

Abstract: In this paper we propose a global optimization-based approach to jointly matching a set of images. The estimated correspondences simultaneously maximize pairwise feature affinities and cycle consistency across multiple images. Unlike previous convex methods relying on semidefinite programming, we formulate the problem as a low-rank matrix recovery problem and show that the desired semidefiniteness of a solution can be spontaneously fulfilled. The low-rank formulation enables us to derive a fast alternating minimization algorithm in order to handle practical problems with thousands of features. Both simulation and real experiments demonstrate that the proposed algorithm can achieve a competitive performance with an order of magnitude speedup compared to the state-of-the-art algorithm. In the end, we demonstrate the applicability of the proposed method to match the images of different object instances and as a result the potential to reconstruct category-specific object models from those images.
Similar papers:
  • 3D Fragment Reassembly Using Integrated Template Guidance and Fracture-Region Matching [pdf] - Kang Zhang, Wuyi Yu, Mary Manhein, Warren Waggenspack, Xin Li
  • Mining And-Or Graphs for Graph Matching and Object Discovery [pdf] - Quanshi Zhang, Ying Nian Wu, Song-Chun Zhu
  • Discrete Tabu Search for Graph Matching [pdf] - Kamil Adamczewski, Yumin Suh, Kyoung Mu Lee
  • Hyperpoints and Fine Vocabularies for Large-Scale Location Recognition [pdf] - Torsten Sattler, Michal Havlena, Filip Radenovic, Konrad Schindler, Marc Pollefeys
  • A Matrix Decomposition Perspective to Multiple Graph Matching [pdf] - Junchi Yan, Hongteng Xu, Hongyuan Zha, Xiaokang Yang, Huanxi Liu, Stephen Chu
Temporal Perception and Prediction in Ego-Centric Video [pdf]
Yipin Zhou, Tamara L. Berg

Abstract: Given a video of an activity, can we predict what will happen next? In this paper we explore two simple tasks related to temporal prediction in egocentric videos of everyday activities. We provide both human experiments to understand how well people can perform on these tasks and computational models for prediction. Experiments indicate that humans and computers can do well on temporal prediction and that personalization to a particular individual or environment provides significantly increased performance. Developing methods for temporal prediction could have far reaching benefits for robots or intelligent agents to anticipate what a person will do, before they do it.
Similar papers:
  • Generating Notifications for Missing Actions: Don't Forget to Turn the Lights Off! [pdf] - Bilge Soran, Ali Farhadi, Linda Shapiro
  • Learning Temporal Embeddings for Complex Video Analysis [pdf] - Vignesh Ramanathan, Kevin Tang, Greg Mori, Li Fei-Fei
  • Unsupervised Semantic Parsing of Video Collections [pdf] - Ozan Sener, Amir R. Zamir, Silvio Savarese, Ashutosh Saxena
  • Describing Videos by Exploiting Temporal Structure [pdf] - Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher Pal, Hugo Larochelle, Aaron Courville
  • Activity Auto-Completion: Predicting Human Activities From Partial Videos [pdf] - Zhen Xu, Laiyun Qing, Jun Miao
Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books [pdf]
Yukun Zhu, Ryan Kiros, Rich Zemel, Ruslan Salakhutdinov, Raquel Urtasun, Antonio Torralba, Sanja Fidler

Abstract: Books are a rich source of both fine-grained information, how a character, an object or a scene looks like, as well as high-level semantics, what someone is thinking, feeling and how these states evolve through a story. This paper aims to align books to their movie releases in order to provide rich descriptive explanations for visual content that go semantically far beyond the captions available in current datasets. To align movies and books we exploit a neural sentence embedding that is trained in an unsupervised way from a large corpus of books, as well as a video-text neural embedding for computing similarities between movie clips and sentences in the book. We propose a context-aware CNN to combine information from multiple sources. We demonstrate good quantitative performance for movie/book alignment and show several qualitative examples that showcase the diversity of tasks our model can be used for.
Similar papers:
  • Learning Like a Child: Fast Novel Visual Concept Learning From Sentence Descriptions of Images [pdf] - Junhua Mao, Xu Wei, Yi Yang, Jiang Wang, Zhiheng Huang, Alan L. Yuille
  • Weakly-Supervised Alignment of Video With Text [pdf] - Piotr Bojanowski, Remi Lajugie, Edouard Grave, Francis Bach, Ivan Laptev, Jean Ponce, Cordelia Schmid
  • Sequence to Sequence - Video to Text [pdf] - Subhashini Venugopalan, Marcus Rohrbach, Jeffrey Donahue, Raymond Mooney, Trevor Darrell, Kate Saenko
  • Guiding the Long-Short Term Memory Model for Image Caption Generation [pdf] - Xu Jia, Efstratios Gavves, Basura Fernando, Tinne Tuytelaars
  • Multimodal Convolutional Neural Networks for Matching Image and Sentence [pdf] - Lin Ma, Zhengdong Lu, Lifeng Shang, Hang Li
Learning a Discriminative Model for the Perception of Realism in Composite Images [pdf]
Jun-Yan Zhu, Philipp Krahenbuhl, Eli Shechtman, Alexei A. Efros

Abstract: What makes an image appear realistic? In this work, we are looking at this question from a data-driven perspective, by learning the perception of visual realism directly from large amounts of unlabeled data. In particular, we train a Convolutional Neural Network (CNN) model that distinguishes natural photographs from automatically generated composite images. The model learns to predict visual realism of a scene in terms of color, lighting and texture compatibility, without any human annotations pertaining to it. Our model outperforms previous works that rely on hand-crafted heuristics for the task of classifying realistic vs. unrealistic photos. Furthermore, we apply our learned model to compute optimal parameters of a compositing method, to maximize the visual realism score predicted by our CNN model. We demonstrate its advantage against existing methods via a human perception study.
Similar papers:
  • Illumination Robust Color Naming via Label Propagation [pdf] - Yuanliu liu, Zejian Yuan, Badong Chen, Jianru Xue, Nanning Zheng
  • Beyond White: Ground Truth Colors for Color Constancy Correction [pdf] - Dongliang Cheng, Brian Price, Scott Cohen, Michael S. Brown
  • Convolutional Color Constancy [pdf] - Jonathan T. Barron
  • Joint Optimization of Segmentation and Color Clustering [pdf] - Ekaterina Lobacheva, Olga Veksler, Yuri Boykov
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
Single Image Pop-Up From Discriminatively Learned Parts [pdf]
Menglong Zhu, Xiaowei Zhou, Kostas Daniilidis

Abstract: We introduce a new approach for estimating a fine grained 3D shape and continuous pose of an object from a single image. Given a training set of view exemplars, we learn and select appearance-based discriminative parts which are mapped onto the 3D model through a facility location optimization. The training set of 3D models is summarized into a set of basis shapes from which we can generalize by linear combination. Given a test image, we detect hypotheses for each part. The main challenge is to select from these hypotheses and compute the 3D pose and shape coefficients at the same time. To achieve this, we optimize a function that considers simultaneously the appearance matching of the parts as well as the geometric reprojection error. We apply the alternating direction method of multipliers (ADMM) to minimize the resulting convex function. Our main and novel contribution is the simultaneous solution for part localization and detailed 3D geometry estimation by maximizing both appearance and geometric compatibility with convex relaxation.
Similar papers:
  • Regressing a 3D Face Shape From a Single Image [pdf] - Sergey Tulyakov, Nicu Sebe
  • Robust Facial Landmark Detection Under Significant Head Poses and Occlusion [pdf] - Yue Wu, Qiang Ji
  • The One Triangle Three Parallelograms Sampling Strategy and Its Application in Shape Regression [pdf] - Mikael Nilsson
  • Regressive Tree Structured Model for Facial Landmark Localization [pdf] - Gee-Sern Hsu, Kai-Hsiang Chang, Shih-Chieh Huang
  • Scene-Domain Active Part Models for Object Representation [pdf] - Zhou Ren, Chaohui Wang, Alan L. Yuille
Learning Ordinal Relationships for Mid-Level Vision [pdf]
Daniel Zoran, Phillip Isola, Dilip Krishnan, William T. Freeman

Abstract: We propose a framework that infers mid-level visual properties of an image by learning about ordinal relationships. Instead of estimating metric quantities directly, the system proposes pairwise relationship estimates for points in the input image. These sparse probabilistic ordinal measurements are globalized to create a dense output map of continuous metric measurements. Estimating order relationships between pairs of points has several advantages over metric estimation: it solves a simpler problem than metric regression; humans are better at relative judgements, so data collection is easier; ordinal relationships are invariant to monotonic transformations of the data, thereby increasing the robustness of the system and providing qualitatively different information. We demonstrate that this framework works well on two important mid-level vision tasks: intrinsic image decomposition and depth from an RGB image. We train two systems with the same architecture on data from these two modalities. We provide an analysis of the resulting models, showing that they learn a number of simple rules to make ordinal decisions. We apply our algorithm to depth estimation, with good results, and intrinsic image decomposition, with state-of-the-art results.
Similar papers:
  • Intrinsic Scene Decomposition From RGB-D images [pdf] - Mohammed Hachama, Bernard Ghanem, Peter Wonka
  • A Comprehensive Multi-Illuminant Dataset for Benchmarking of the Intrinsic Image Algorithms [pdf] - Shida Beigpour, Andreas Kolb, Sven Kunz
  • Intrinsic Depth: Improving Depth Transfer With Intrinsic Images [pdf] - Naejin Kong, Michael J. Black
  • Intrinsic Decomposition of Image Sequences From Local Temporal Variations [pdf] - Pierre-Yves Laffont, Jean-Charles Bazin
  • Learning Data-Driven Reflectance Priors for Intrinsic Image Decomposition [pdf] - Tinghui Zhou, Philipp Krahenbuhl, Alexei A. Efros
HARF: Hierarchy-Associated Rich Features for Salient Object Detection [pdf]
Wenbin Zou, Nikos Komodakis

Abstract: The state-of-the-art salient object detection models are able to perform well for relatively simple scenes, yet for more complex ones, they still have difficulties in highlighting salient objects completely from background, largely due to the lack of sufficiently robust features for saliency prediction. To address such an issue, this paper proposes a novel hierarchy-associated feature construction framework for salient object detection, which is based on integrating elementary features from multi-level regions in a hierarchy. Furthermore, multi-layered deep learning features are introduced and incorporated as elementary features into this framework through a compact integration scheme. This leads to a rich feature representation, which is able to represent the context of the whole object/background and is much more discriminative as well as robust for salient object detection. Extensive experiments on the most widely used and challenging benchmark datasets demonstrate that the proposed approach substantially outperforms the state-of-theart on salient object detection.
Similar papers:
  • A Data-Driven Metric for Comprehensive Evaluation of Saliency Models [pdf] - Jia Li, Changqun Xia, Yafei Song, Shu Fang, Xiaowu Chen
  • Generic Promotion of Diffusion-Based Salient Object Detection [pdf] - Peng Jiang, Nuno Vasconcelos, Jingliang Peng
  • Cluster-Based Point Set Saliency [pdf] - Flora Ponjou Tasse, Jiri Kosinka, Neil Dodgson
  • SALICON: Reducing the Semantic Gap in Saliency Prediction by Adapting Deep Neural Networks [pdf] - Xun Huang, Chengyao Shen, Xavier Boix, Qi Zhao
  • Minimum Barrier Salient Object Detection at 80 FPS [pdf] - Jianming Zhang, Stan Sclaroff, Zhe Lin, Xiaohui Shen, Brian Price, Radomir Mech
Video Matting via Sparse and Low-Rank Representation [pdf]
Dongqing Zou, Xiaowu Chen, Guangying Cao, Xiaogang Wang

Abstract: We introduce a novel method of video matting via sparse and low-rank representation. Previous matting methods [10, 9] introduced a nonlocal prior to estimate the alpha matte and have achieved impressive results on some data. However, on one hand, searching inadequate or excessive samples may miss good samples or introduce noise; on the other hand, it is difficult to construct consistent nonlocal structures for pixels with similar features, yielding spatially and temporally inconsistent video mattes. In this paper, we proposed a novel video matting method to achieve spatially and temporally consistent matting result. Toward this end, a sparse and low-rank representation model is introduced to pursue consistent nonlocal structures for pixels with similar features. The sparse representation is used to adaptively select best samples and accurately construct the nonlocal structures for all pixels, while the low-rank representation is used to globally ensure consistent nonlocal structures for pixels with similar features. The two representations are combined to generate consistent video mattes. Experimental results show that our method has achieved high quality results in a variety of challenging examples featuring illumination changes, feature ambiguity, topology changes, transparency variation, dis-occlusion, fast motion and motion blur.
Similar papers:
  • Dynamic Texture Recognition via Orthogonal Tensor Dictionary Learning [pdf] - Yuhui Quan, Yan Huang, Hui Ji
  • Co-Interest Person Detection From Multiple Wearable Camera Videos [pdf] - Yuewei Lin, Kareem Abdelfatah, Youjie Zhou, Xiaochuan Fan, Hongkai Yu, Hui Qian, Song Wang
  • Frequency-Based Environment Matting by Compressive Sensing [pdf] - Yiming Qian, Minglun Gong, Yee-Hong Yang
  • Conformal and Low-Rank Sparse Representation for Image Restoration [pdf] - Jianwei Li, Xiaowu Chen, Dongqing Zou, Bo Gao, Wei Teng
  • Image Matting With KL-Divergence Based Sparse Sampling [pdf] - Levent Karacan, Aykut Erdem, Erkut Erdem
Interactive Visual Hull Refinement for Specular and Transparent Object Surface Reconstruction [pdf]
Xinxin Zuo, Chao Du, Sen Wang, Jiangbin Zheng, Ruigang Yang

Abstract: In this paper we present a method of using standard multi-view images for 3D surface reconstruction of nonLambertian objects. We extend the original visual hull concept to incorporate 3D cues presented by internal occluding contours, i.e., occluding contours that are inside the objects silhouettes. We discovered that these internal contours, which are results of convex parts on an objects surface, can lead to a tighter fit than the original visual hull. We formulated a new visual hull refinement scheme Locally Convex Carving that can completely reconstruct concavity caused by two or more intersecting convex surfaces. In addition we develop a novel approach for contour tracking given labeled contours in sparse key frames. It is designed specifically for highly specular or transparent objects, for which assumptions made in traditional contour detection/tracking methods, such as highest gradient and stationary texture edges, are no longer valid. It is formulated as an energy minimization function where several novel terms are developed to increase robustness. Based on the two core algorithms, we have developed an interactive system for 3D modeling. We have validated our system, both quantitatively and qualitatively, with four datasets of different object materials. Results show that we are able to generate visually pleasing models for very challenging cases.
Similar papers:
  • Contour Guided Hierarchical Model for Shape Matching [pdf] - Yuanqi Su, Yuehu Liu, Bonan Cuan, Nanning Zheng
  • Robust Image Segmentation Using Contour-Guided Color Palettes [pdf] - Xiang Fu, Chien-Yi Wang, Chen Chen, Changhu Wang, C.-C. Jay Kuo
  • Contour Box: Rejecting Object Proposals Without Explicit Closed Contours [pdf] - Cewu Lu, Shu Liu, Jiaya Jia, Chi-Keung Tang
  • StereoSnakes: Contour Based Consistent Object Extraction For Stereo Images [pdf] - Ran Ju, Tongwei Ren, Gangshan Wu
  • Contour Flow: Middle-Level Motion Estimation by Combining Motion Segmentation and Contour Alignment [pdf] - Huijun Di, Qingxuan Shi, Feng Lv, Ming Qin, Yao Lu