Difference between revisions of "Main Page"

From cvss
 
(339 intermediate revisions by 9 users not shown)
Line 1: Line 1:
<Big>'''Computer Vision Student Seminar'''</Big>
+
<Big>'''Computer Vision Student Seminars'''</Big>
  
The Computer Vision Student Seminar at the University of Maryland College Park is a student-run series of talks given by [http://www.cfar.umd.edu/cvl/meetthe.html#Graduate current graduate students] for [http://www.cfar.umd.edu/cvl/meetthe.html#Graduate current graduate students].
+
The Computer Vision Student Seminars at the University of Maryland College Park are a student-run series of talks given by [http://www.cfar.umd.edu/cvl/meetthe.html#Graduate current graduate students] for [http://www.cfar.umd.edu/cvl/meetthe.html#Graduate current graduate students].
  
 +
To receive regular information about the Computer Vision Student Seminars, subscribe to our [https://mailman.cs.umd.edu/mailman/listinfo/cvss mailing list] or our [http://talks.cs.umd.edu/lists/12 talks list].
  
 
==Description==
 
==Description==
Line 13: Line 14:
 
* Provide speaking opportunities for computer vision students.
 
* Provide speaking opportunities for computer vision students.
  
 
+
The guidelines for the format are:
==Format==
 
  
 
* An hour-long weekly meeting, consisting of one 20-40 minute talk followed by discussion and food.   
 
* An hour-long weekly meeting, consisting of one 20-40 minute talk followed by discussion and food.   
Line 20: Line 20:
 
* Topics may include current research, past research, general topic presentations, paper summaries and critiques, or anything else beneficial to the computer vision graduate student community.
 
* Topics may include current research, past research, general topic presentations, paper summaries and critiques, or anything else beneficial to the computer vision graduate student community.
  
 +
==Schedule Fall 2015==
 +
 +
All talks take place on Thursdays at 3:30pm in AVW 3450.
 +
 +
{| class="wikitable" cellpadding="10" border="1" cellspacing="1"
 +
|-
 +
! Date
 +
! Speaker
 +
! Title
 +
|-
 +
| December 3
 +
| Angjoo Kanazawa
 +
| Learning 3D Deformation of Animals from 2D Images
 +
|-
 +
| December 10
 +
| Xintong Han
 +
| Automated Event Retrieval using Web Trained Detectors
 +
|}
 +
 +
==Talk Abstracts Spring 2015==
 +
 +
 +
===Learning 3D Deformation of Animals from 2D Images===
 +
Speaker: [http://www.umiacs.umd.edu/~kanazawa/ Angjoo Kanazawa] -- Date: December 3, 2015
 +
 +
Abstract: Understanding how an animal can deform and articulate is essential for a realistic modification of its 3D model. In this paper, we show that such information can be learned from user-clicked 2D images and a template 3D model of the target animal. We present a volumetric deformation framework that produces a set of new 3D models by deforming a template 3D model according to a set of user-clicked images. Our framework is based on a novel locally-bounded deformation energy, where every local region has its own stiffness value that bounds how much distortion is allowed at that location. We jointly learn the local stiffness bounds as we deform the template 3D mesh to match each user-clicked image. We show that this seemingly complex task can be solved as a sequence of convex optimization problems. We demonstrate the effectiveness of our approach on cats and horses, which are highly deformable and articulated animals. Our framework produces new 3D models of animals that are significantly more plausible than methods without learned stiffness.
 +
 +
Link: [http://arxiv.org/pdf/1507.07646v1.pdf paper]
 +
 +
===Automated Event Retrieval using Web Trained Detectors===
 +
 +
Speaker: [http://www.umiacs.umd.edu/~xintong/ Xintong Han] -- Date: December 10, 2015
 +
 +
Abstract: Complex event retrieval is a challenging research problem, especially when no training videos are available. An alternative to collecting training videos is to train a large semantic concept bank a priori. Given a text description of an event, event retrieval is performed by selecting concepts linguistically related to the event description and fusing the concept responses on unseen videos. However, defining an exhaustive concept lexicon and pre-training it requires vast computational resources. Therefore, recent approaches automate concept discovery and training by leveraging large amounts of weakly annotated web data. Compact visually salient concepts are automatically obtained by the use of concept pairs or, more generally, n-grams. However, not all visually salient n-grams are necessarily useful for an event query - some combinations of concepts may be visually compact but irrelevant--and this drastically affects performance. We propose an event retrieval algorithm that constructs pairs of automatically discovered concepts and then prunes those concepts that are unlikely to be helpful for retrieval. Pruning depends both on the query and on the specific video instance being evaluated. Our approach also addresses calibration and domain adaptation issues that arise when applying concept detectors to unseen videos. We demonstrate large improvements over other vision based systems on the TRECVID MED 13 dataset.
  
==Subscribe to the Mailing List==
+
Link: [http://arxiv.org/pdf/1509.07845v1.pdf paper]
  
To receive regular information about the Computer Vision Student Seminar, subscribe to the mailing list by following the instructions [https://mailman.cs.umd.edu/mailman/listinfo/cvss here].
+
==Past Semesters==
 +
* [[Cvss:Spring2015| Spring 2015]]
 +
* [[cvss fall2014|Fall 2014]]
 +
* [[cvss_spring2014|Spring 2014]]
 +
* [[cvss_fall2013|Fall 2013]]
 +
* [[cvss_summer2013|Summer 2013]]
 +
* [[cvss_spring2013|Spring 2013]]
 +
* [[cvss_fall2012|Fall 2012]]
 +
* [[cvss_spring2012|Spring 2012]]
 +
* [[cvss_fall2011|Fall 2011]]
 +
* [[cvss_summer2011|Summer 2011]]
  
 +
==Funded By==
 +
* Computer Vision Faculty
 +
<!-- * '''[http://www.northropgrumman.com/ Northrop Grumman]''' -->
  
 
==Current Seminar Series Coordinators==
 
==Current Seminar Series Coordinators==
Line 30: Line 77:
 
Emails are at umiacs.umd.edu.
 
Emails are at umiacs.umd.edu.
  
{| class="wikitable" cellpadding="5"
+
{| cellpadding="1"
 
|-
 
|-
| Anne Jorstad, jorstad@
+
| [http://sites.google.com/site/austinomyers/ Austin Myers], amyers@
| (student of [http://www.cs.umd.edu/~djacobs/ Professor David Jacobs])
+
| (student of [http://www.cfar.umd.edu/~yiannis/ Professor Yiannis Aloimonos])
 
|-
 
|-
| Sameh Khamis, sameh@
+
| [http://www.umiacs.umd.edu/~kanazawa/ Angjoo Kanazawa], kanazawa@
 +
| (student of [http://cs.umd.edu/~djacobs/ Professor David Jacobs])
 +
|-
 +
| [http://sites.google.com/site/yechengxi/ Chenxi Ye] cxy@
 +
| (student of [http://www.cfar.umd.edu/~yiannis/ Professor Yiannis Aloimonos])
 +
|-
 +
| [http://www.umiacs.umd.edu/~xintong/ Xintong Han], xintong@
 
| (student of [http://www.umiacs.umd.edu/~lsd/ Professor Larry Davis])
 
| (student of [http://www.umiacs.umd.edu/~lsd/ Professor Larry Davis])
 
|-
 
|-
| Sima Taheri, taheri@
+
| [http://www.cs.umd.edu/~bharat/ Bharat Singh], bharat@
| (student of [http://www.umiacs.umd.edu/~rama/ Professor Rama Chellappa])
+
| (student of [http://www.umiacs.umd.edu/~lsd/ Professor Larry Davis])
 
|-
 
|-
| Ching Lik Teo, cteo@
+
| [http://bcsiriuschen.github.io/ Bor-Chun (Sirius) Chen], sirius@
| (student of [http://www.cfar.umd.edu/~yiannis/ Professor Yiannis Aloimonos])
+
| (student of [http://www.umiacs.umd.edu/~lsd/ Professor Larry Davis])
 
|}
 
|}
  
 
+
Gone but not forgotten.
==Schedule Fall 2011==
+
{| cellpadding="1"
 
 
All talks take place Thursdays at 4pm in AVW 3450.
 
 
 
{| class="wikitable" cellpadding="10" border="1" cellspacing="0"
 
 
|-
 
|-
! Date
+
| [http://www.umiacs.umd.edu/~jhchoi/ Jonghyun Choi], jhchoi@
! Speaker
+
| (student of [http://www.umiacs.umd.edu/~lsd/ Professor Larry Davis])
! Title
 
 
|-
 
|-
| September 8
+
| Ching-Hui Chen, ching@
| Vishal Patel
+
| (student of [http://www.umiacs.umd.edu/~rama/ Professor Rama Chellappa])
| Wavelets with Composite Dilations
 
|-
 
| September 15
 
|
 
 
|
 
|
 
|-
 
|-
| September 22
+
| [http://ravitejav.weebly.com/ Raviteja Vemulapalli], raviteja @
| Dave Shaw
+
| (student of [http://www.umiacs.umd.edu/~rama/ Professor Rama Chellappa])
|
 
 
|-
 
|-
| September 29
+
| [http://www.umiacs.umd.edu/~sameh/ Sameh Khamis]
 
|  
 
|  
|
 
|-
 
| October 6
 
|
 
|
 
 
|-
 
|-
| October 13
+
| [http://www.umiacs.umd.edu/~ejaz/ Ejaz Ahmed]
 
|  
 
|  
|
 
 
|-
 
|-
| October 20
+
| [http://cvlabwww.epfl.ch/~jorstad/ Anne Jorstad]
|
+
| now at EPFL
|
 
 
|-
 
|-
| October 27
+
| [http://www.umiacs.umd.edu/~jni/ Jie Ni]
| Garrett Warnell
+
| now at Sony
|
 
 
|-
 
|-
| November 3
+
| [http://www.umiacs.umd.edu/~taheri/ Sima Taheri]
| Abhishek Sharma
+
|  
|
 
 
|-
 
|-
| November 10
+
| [http://www.umiacs.umd.edu/~cteo/ Ching Lik Teo]
| (ICCV, meeting TBD)
+
|  
|
 
|-
 
| November 17
 
| (no meeting, CVPR deadline 11/21)
 
|
 
|-
 
| November 24
 
| (no meeting, Thanksgiving)
 
|
 
|-
 
| December 1
 
| Nitesh Shroff
 
|
 
|-
 
| December 8
 
| Ming-Yu Liu
 
|
 
|-
 
| December 15
 
| (no meeting, final exams)
 
|
 
 
|}
 
|}
 
 
==Talk Abstracts Fall 2011==
 
 
===Wavelets with Composite Dilations===
 
Speaker: Vishal Patel -- Date: September 8, 2011
 
 
Sparse representation of visual information lies at the foundation of many image processing applications, such as image restoration and compression. It is well known that wavelets provide a very sparse representation for a large class of signals and images. For instance, from a continuous perspective, wavelets can be shown to sparsely represent one-dimensional signals that are smooth away from point discontinuities. Unfortunately, separable wavelet transforms have some limitations in higher dimensions. For this reason, in recent years there has been considerable interest in obtaining directionally-oriented image decompositions. Wavelets with composite dilations offer a general and especially effective framework for the construction of such representations.  In this talk, I will discuss the theory and implementation of several recently introduced multiscale directional transforms. Then, I will present a new general scheme for creating an M-channel directional filter bank. An advantage of an M-channel directional filter bank is that it can project the image directly onto the desired basis. Applications in image denoising, deconvolution and image enhancement will be presented.
 
 
==Schedule Summer 2011==
 
[[cvss_summer2011|Schedule Summer 2011]]
 
 
 
== Wiki Editing ==
 
 
Consult the [http://meta.wikimedia.org/wiki/Help:Contents User's Guide] for information on using the wiki software.
 
 
* [http://www.mediawiki.org/wiki/Help:Configuration_settings Configuration settings list]
 
* [http://www.mediawiki.org/wiki/Help:FAQ MediaWiki FAQ]
 
* [http://mail.wikimedia.org/mailman/listinfo/mediawiki-announce MediaWiki release mailing list]
 

Latest revision as of 23:40, 3 December 2015

Computer Vision Student Seminars

The Computer Vision Student Seminars at the University of Maryland College Park are a student-run series of talks given by current graduate students for current graduate students.

To receive regular information about the Computer Vision Student Seminars, subscribe to our mailing list or our talks list.

Description[edit]

The purpose of these talks is to:

  • Encourage interaction between computer vision students;
  • Provide an opportunity for computer vision students to be aware of and possibly get involved in the research their peers are conducting;
  • Provide an opportunity for computer vision students to receive feedback on their current research;
  • Provide speaking opportunities for computer vision students.

The guidelines for the format are:

  • An hour-long weekly meeting, consisting of one 20-40 minute talk followed by discussion and food.
  • The talks are meant to be casual and discussion is encouraged.
  • Topics may include current research, past research, general topic presentations, paper summaries and critiques, or anything else beneficial to the computer vision graduate student community.

Schedule Fall 2015[edit]

All talks take place on Thursdays at 3:30pm in AVW 3450.

Date Speaker Title
December 3 Angjoo Kanazawa Learning 3D Deformation of Animals from 2D Images
December 10 Xintong Han Automated Event Retrieval using Web Trained Detectors

Talk Abstracts Spring 2015[edit]

Learning 3D Deformation of Animals from 2D Images[edit]

Speaker: Angjoo Kanazawa -- Date: December 3, 2015

Abstract: Understanding how an animal can deform and articulate is essential for a realistic modification of its 3D model. In this paper, we show that such information can be learned from user-clicked 2D images and a template 3D model of the target animal. We present a volumetric deformation framework that produces a set of new 3D models by deforming a template 3D model according to a set of user-clicked images. Our framework is based on a novel locally-bounded deformation energy, where every local region has its own stiffness value that bounds how much distortion is allowed at that location. We jointly learn the local stiffness bounds as we deform the template 3D mesh to match each user-clicked image. We show that this seemingly complex task can be solved as a sequence of convex optimization problems. We demonstrate the effectiveness of our approach on cats and horses, which are highly deformable and articulated animals. Our framework produces new 3D models of animals that are significantly more plausible than methods without learned stiffness.

Link: paper

Automated Event Retrieval using Web Trained Detectors[edit]

Speaker: Xintong Han -- Date: December 10, 2015

Abstract: Complex event retrieval is a challenging research problem, especially when no training videos are available. An alternative to collecting training videos is to train a large semantic concept bank a priori. Given a text description of an event, event retrieval is performed by selecting concepts linguistically related to the event description and fusing the concept responses on unseen videos. However, defining an exhaustive concept lexicon and pre-training it requires vast computational resources. Therefore, recent approaches automate concept discovery and training by leveraging large amounts of weakly annotated web data. Compact visually salient concepts are automatically obtained by the use of concept pairs or, more generally, n-grams. However, not all visually salient n-grams are necessarily useful for an event query - some combinations of concepts may be visually compact but irrelevant--and this drastically affects performance. We propose an event retrieval algorithm that constructs pairs of automatically discovered concepts and then prunes those concepts that are unlikely to be helpful for retrieval. Pruning depends both on the query and on the specific video instance being evaluated. Our approach also addresses calibration and domain adaptation issues that arise when applying concept detectors to unseen videos. We demonstrate large improvements over other vision based systems on the TRECVID MED 13 dataset.

Link: paper

Past Semesters[edit]

Funded By[edit]

  • Computer Vision Faculty

Current Seminar Series Coordinators[edit]

Emails are at umiacs.umd.edu.

Austin Myers, amyers@ (student of Professor Yiannis Aloimonos)
Angjoo Kanazawa, kanazawa@ (student of Professor David Jacobs)
Chenxi Ye cxy@ (student of Professor Yiannis Aloimonos)
Xintong Han, xintong@ (student of Professor Larry Davis)
Bharat Singh, bharat@ (student of Professor Larry Davis)
Bor-Chun (Sirius) Chen, sirius@ (student of Professor Larry Davis)

Gone but not forgotten.

Jonghyun Choi, jhchoi@ (student of Professor Larry Davis)
Ching-Hui Chen, ching@ (student of Professor Rama Chellappa)
Raviteja Vemulapalli, raviteja @ (student of Professor Rama Chellappa)
Sameh Khamis
Ejaz Ahmed
Anne Jorstad now at EPFL
Jie Ni now at Sony
Sima Taheri
Ching Lik Teo