{"188481":{"#nid":"188481","#data":{"type":"event","title":"Robotics and Intelligent Machines Center (RIM) Seminar-Kristen Grauman","body":[{"value":"\u003Cp\u003EUnless otherwise noted, all seminars are held in room 1116 in the Marcus Nanotechnology Building from 12-1 p.m. Seminars are open to the public.\u003C\/p\u003E\u003Cp class=\u0022p1\u0022\u003EKristen Grauman, associate professor of computer science at UT-Austin, presents \u0022Visual Search and Summarization\u0022 as part of the RIM Seminar Series.\u003C\/p\u003E\u003Cp class=\u0022p1\u0022\u003E\u003Cstrong\u003EAbstract\u003Cbr \/\u003E\u003C\/strong\u003EWidespread visual sensors and unprecedented connectivity have left us awash with visual data--from online photo collections, home videos, news footage, medical images, or surveillance feeds. How can we efficiently browse image and video collections based on semantically meaningful criteria? How can we bring order to the data, beyond manually defined keyword tags? We are exploring these questions in our recent work on interactive visual search and summarization.\u003C\/p\u003E\u003Cp\u003EI will first present a novel form of interactive feedback for visual search, in which a user helps pinpoint the content of interest by making visual comparisons between his envisioned target and reference images. The approach relies on a powerful mid-level representation of interpretable relative attributes to connect the user\u2019s descriptions to the system\u2019s internal features. Whereas traditional feedback limits input to coarse binary labels, the proposed \u201cWhittleSearch\u201d lets a user state precisely what about an image is relevant, leading to more rapid convergence to the desired content. Turning to issues in video browsing, I will then present our work on automatic summarization of egocentric videos. Given a long video captured with a wearable camera, our method produces a short storyboard summary. Whereas existing summarization methods define sampling-based objectives (e.g., to maximize diversity in the output summary), we take a \u201cstory-driven\u201d approach that predicts the high-level importance of objects and their influence between subevents. We show this leads to substantially more accurate summaries, allowing a viewer to quickly understand the gist of a long video. This work is being conducted with Adriana Kovashka, Yong Jae Lee, Devi Parikh, and Lu Zheng.\u003C\/p\u003E","summary":null,"format":"limited_html"}],"field_subtitle":"","field_summary":[{"value":"\u003Cp\u003EUnless otherwise noted, all seminars are held in room 1116 in the Marcus Nanotechnology Building from 12-1 p.m. Seminars are open to the public.\u003C\/p\u003E\u003Cp class=\u0022p1\u0022\u003EKristen Grauman, associate professor of computer science at UT-Austin, presents \u0022Visual Search and Summarization\u0022 as part of the RIM Seminar Series.\u003C\/p\u003E","format":"limited_html"}],"field_summary_sentence":[{"value":"RIM seminar by Kristen Grauman of the University of Texas at Austin"}],"uid":"27255","created_gmt":"2013-01-31 12:08:59","changed_gmt":"2016-10-08 02:02:24","author":"Josie Giles","boilerplate_text":"","field_publication":"","field_article_url":"","field_event_time":{"event_time_start":"2013-03-06T11:00:00-05:00","event_time_end":"2013-03-06T12:00:00-05:00","event_time_end_last":"2013-03-06T12:00:00-05:00","gmt_time_start":"2013-03-06 16:00:00","gmt_time_end":"2013-03-06 17:00:00","gmt_time_end_last":"2013-03-06 17:00:00","rrule":null,"timezone":"America\/New_York"},"extras":[],"hg_media":{"188451":{"id":"188451","type":"image","title":"Kristen Grauman","body":null,"created":"1449179101","gmt_created":"2015-12-03 21:45:01","changed":"1475894611","gmt_changed":"2016-10-08 02:43:31","alt":"Kristen Grauman","file":{"fid":"196216","name":"kristen-grauman.jpg","image_path":"\/sites\/default\/files\/images\/kristen-grauman_0.jpg","image_full_path":"http:\/\/www.tlwarc.hg.gatech.edu\/\/sites\/default\/files\/images\/kristen-grauman_0.jpg","mime":"image\/jpeg","size":56209,"path_740":"http:\/\/www.tlwarc.hg.gatech.edu\/sites\/default\/files\/styles\/740xx_scale\/public\/images\/kristen-grauman_0.jpg?itok=XYaom67l"}}},"media_ids":["188451"],"related_links":[{"url":"http:\/\/www.cs.utexas.edu\/~grauman\/","title":"Kristen Grauman"},{"url":"http:\/\/www.robotics.gatech.edu\/","title":"Robotics at Georgia Tech"}],"groups":[{"id":"47223","name":"College of Computing"},{"id":"142761","name":"IRIM"}],"categories":[],"keywords":[{"id":"57331","name":"Kristen Grauman"},{"id":"667","name":"robotics"},{"id":"40311","name":"Robotics \u0026 Intelligent Machines (RIM)"}],"core_research_areas":[],"news_room_topics":[],"event_categories":[{"id":"1795","name":"Seminar\/Lecture\/Colloquium"}],"invited_audience":[],"affiliations":[],"classification":[],"areas_of_expertise":[],"news_and_recent_appearances":[],"phone":[],"contact":[{"value":"\u003Cp\u003EJosie Giles\u003Cbr \/\u003ERIM Communications Officer\u003Cbr \/\u003E\u003Ca href=\u0022mailto:josie@gatech.edu\u0022\u003Ejosie@gatech.edu\u003C\/a\u003E\u003C\/p\u003E","format":"limited_html"}],"email":[],"slides":[],"orientation":[],"userdata":""}}}