Main Page
Computer Vision Student Seminars
The Computer Vision Student Seminars at the University of Maryland College Park are a student-run series of talks given by current graduate students for current graduate students.
To receive regular information about the Computer Vision Student Seminars, subscribe to the mailing list by following the instructions here.
Description[edit]
The purpose of these talks is to:
- Encourage interaction between computer vision students;
- Provide an opportunity for computer vision students to be aware of and possibly get involved in the research their peers are conducting;
- Provide an opportunity for computer vision students to receive feedback on their current research;
- Provide speaking opportunities for computer vision students.
The guidelines for the format are:
- An hour-long weekly meeting, consisting of one 20-40 minute talk followed by discussion and food.
- The talks are meant to be casual and discussion is encouraged.
- Topics may include current research, past research, general topic presentations, paper summaries and critiques, or anything else beneficial to the computer vision graduate student community.
Schedule Fall 2011[edit]
All talks take place Thursdays at 4pm in AVW 3450.
Date | Speaker | Title |
---|---|---|
September 8 | Vishal Patel | Wavelets with Composite Dilations |
September 15 | Radu Dondera | Kernel PLS Regression for Robust Monocular Pose Estimation |
September 22 | Dave Shaw | Regularization and Localization for Prediction on Manifolds |
September 29 | Douglas Summers-Stay (room 3165) | Scene Classification with Visual Filters |
October 6 | Arpit Jain | Learning What and How of Contextual Models for Scene Labeling |
October 13 | Yi-Chen Chen | Rotation Invariant Simultaneous Clustering and Dictionary Learning |
October 20 | Anne Jorstad | Deformation and Lighting Insensitive Face Recognition: From Optical Flow to Geodesics |
October 27 | Garrett Warnell | Compressive Sensing in Visual Tracking |
November 3 | Abhishek Sharma | Cross-modal classification and retrieval : Techniques and challenges |
November 10 | Cheuk Yiu Ip | Saliency-Assisted Navigation of Very Large Landscape Images |
November 17 | (no meeting, CVPR deadline 11/21) | |
November 24 | (no meeting, Thanksgiving) | |
December 1 | Nitesh Shroff | |
December 8 | Ming-Yu Liu | |
December 15 | (no meeting, final exams) |
Talk Abstracts Fall 2011[edit]
Wavelets with Composite Dilations[edit]
Speaker: Vishal Patel -- Date: September 8, 2011
Sparse representation of visual information lies at the foundation of many image processing applications, such as image restoration and compression. It is well known that wavelets provide a very sparse representation for a large class of signals and images. For instance, from a continuous perspective, wavelets can be shown to sparsely represent one-dimensional signals that are smooth away from point discontinuities. Unfortunately, separable wavelet transforms have some limitations in higher dimensions. For this reason, in recent years there has been considerable interest in obtaining directionally-oriented image decompositions. Wavelets with composite dilations offer a general and especially effective framework for the construction of such representations. In this talk, I will discuss the theory and implementation of several recently introduced multiscale directional transforms. Then, I will present a new general scheme for creating an M-channel directional filter bank. An advantage of an M-channel directional filter bank is that it can project the image directly onto the desired basis. Applications in image denoising, deconvolution and image enhancement will be presented.
Kernel PLS Regression for Robust Monocular Pose Estimation[edit]
Speaker: Radu Dondera -- Date: September 15, 2011
We evaluate the robustness of five regression techniques for monocular 3D pose estimation. While most of the discriminative pose estimation methods focus on overcoming the fundamental problem of insufficient training data, we are interested in characterizing performance improvement for increasingly large training sets. Commercially available rendering software allows us to efficiently generate large numbers of realistic images of poses from diverse actions. Inspired by recent work in human detection, we apply PLS and kPLS regression to pose estimation. We observe that kPLS regression incrementally approximates GP regression using the strongest nonlinear correlations between image features and pose. This provides robustness, and our experiments show kPLS regression is more robust than two GP-based state-of-the-art methods for pose estimation. We address the ambiguity problem of pose estimation by random partitioning of the pose space and report results on the HumanEva dataset.
Regularization and Localization for Prediction on Manifolds[edit]
Speaker: David Shaw -- Date: September 22, 2011
In data analysis, one is interested in using the information about the response variable contained in the predictors in the best way possible. This can lead to problems when the predictors are highly collinear, as it implies an inherent lower-dimensional structure in the data. One method of analyzing data of this form is to make the assumption that these structured dependencies arise due to the predictors lying on some implicit lower-dimensional manifold. This assumption helps solve the problem of reducing the dimension of the predictors in the interest of removing some redundant information, but it introduces the problem of analyzing the transformed data. In particular, making accurate predictions with the lower-dimensional data that can be interpreted in the higher-dimensional space can be difficult. The technique of weighted regression with regularization on the model parameters can help to overcome these issues.
Scene Classification with Visual Filters[edit]
Speaker: Douglas Summers-Stay -- Date: September 29, 2011
"Scene Classification" is the computer vision problem of labeling all the pixels in an image according to the class they fall into, such as "street," "tree," or "person." A tool we have developed here at the computer vision lab called "visual filters" uses a series of nonlinear filters to attempt to create such classification maps. I will discuss what we are doing now and how we can incorporate ideas from "deep learning" to improve this in the future. An introduction for beginners with some examples is here.
Learning What and How of Contextual Models for Scene Labeling[edit]
Speaker: Arpit Jain -- Date: October 6, 2011
In this talk I will discuss about a data-driven approach to predict the importance of edges and construct a Markov network for image analysis based on statistical models of global and local image features. Most of the previous approaches used either a fixed fully connected Markov Network(MN) or ad-hoc neighborhood connected MN. But not edges in MN are useful and this is what I will show during my talk. I will also address the coupled problem of predicting the feature weights associated with each edge of a Markov network for evaluation of context. Experimental results indicate that this scene dependent structure construction model eliminates spurious edges and improves performance over fully-connected and neighborhood connected Markov network.
Rotation Invariant Simultaneous Clustering and Dictionary Learning[edit]
Speaker: Yi-Chen Chen -- Date: October 13, 2011
We present an approach that simultaneously clusters database members and learns dictionaries from the clusters. The method learns dictionaries in the Radon transform domain, while clustering in the image domain. The main feature of the proposed approach is that it provides rotation invariant clustering which is useful in Content Based Image Retrieval (CBIR). We demonstrate through experimental results that the proposed rotation invariant clustering provides good retrieval performance than the standard Gabor-based method that has similar objectives.
Deformation and Lighting Insensitive Face Recognition: From Optical Flow to Geodesics[edit]
Speaker: Anne Jorstad -- Date: October 20, 2011
We seek to solve the face identification problem across variations in expression and lighting together in a single framework. In order to understand variations in expression, a dense correspondence between images must be found, leading to algorithms similar to Optical Flow. We present a new lighting-insensitive metric to drive this Optical Flow-like framework. An extension of this work to the manifold of face images is then proposed, where a curve on the manifold represents the way a face might morph through time, allowing pixels to vary slowly as properties of the face change. The length of the geodesic connecting a pair of faces defines their similarity for nearest neighbor matching.
Compressive Sensing in Visual Tracking[edit]
Speaker: Garrett Warnell -- Date: October 27, 2011
Visual tracking is a classical computer vision task. However, the ubiquity of modern sensors makes it more difficult due to the large amount of data available for processing. The emerging theory of compressive sensing has the potential to address this problem in that it promises the ability to reduce the amount of data collected without sacrificing the amount of information within. In this talk, I will review recent research research toward the adaptation of some computer vision algorithms commonly used in visual tracking such that they can operate in the lower-dimensional compressive domain. Specifically, background subtraction and particle filtering will be discussed.
Cross-modal classification and retrieval : Techniques and challenges[edit]
Speaker: Abhishek Sharma -- Date: November 3, 2011
Classification data arrives in multiple forms of representations and distributions (modality) having a common underlying content. Classification or Retrieval is required to be done solely based on the content irrespective of the modality. For example - given a text description of a topic (history) find appropriate images from a database OR given a person's face image in some pose and lighting which is different than that of the gallery, find the matching face OR based on user supplied tags find matching images from the database. These problems are finding applications everywhere because of wide-spread Internet and extremely cheap sensors (cameras and keyboards). In this talk, we will go over some popular techniques from the literature to tackle the problem of cross-modal classification and retrieval. Specifically, I will be discussing Canonical Correlational Analysis (and variants), Partial Least Square, Bilinear Model (Freeman and Tannenbaum), Tied Factor Analysis, Probabilstic LDA, Multi-view LDA and SVM-2k along with detailed pros and cons of each of these. Then I will present a comparative application of all these approaches along with recent methods for pose and lighting invariant face recognition as a case study.
Past Semesters[edit]
Current Seminar Series Coordinators[edit]
Emails are at umiacs.umd.edu.
Anne Jorstad, jorstad@ | (student of Professor David Jacobs) |
Sameh Khamis, sameh@ | (student of Professor Larry Davis) |
Sima Taheri, taheri@ | (student of Professor Rama Chellappa) |
Ching Lik Teo, cteo@ | (student of Professor Yiannis Aloimonos) |