{"490711":{"#nid":"490711","#data":{"type":"event","title":"Seminar - Weijie Su","body":[{"value":"\u003Cp\u003ETITLE:\u0026nbsp; Multiple Testing and Adaptive Estimation via the Sorted L-One Norm\u003C\/p\u003E\u003Cp\u003EABSTRACT:\u003C\/p\u003E\u003Cp\u003EIn many real-world statistical problems, we observe a large number of potentially explanatory variables of which a majority may be irrelevant. For this type of problem, controlling the false discovery rate (FDR) guarantees that most of the discoveries are truly explanatory and thus replicable. In this talk, we propose a new method named SLOPE to control the FDR in sparse high-dimensional linear regression. This computationally efficient procedure works by regularizing the fitted coefficients according to their ranks: the higher the rank, the larger the penalty. This is analogous to the Benjamini-Hochberg procedure, which compares more significant p-values with more stringent thresholds. Whenever the columns of the design matrix are not strongly correlated, we show empirically that SLOPE obtains FDR control at a reasonable level while offering substantial power.\u0026nbsp;\u003C\/p\u003E\u003Cp\u003E\u0026nbsp;\u003C\/p\u003E\u003Cp\u003EAlthough SLOPE is developed from a multiple testing viewpoint, we show the surprising result that it achieves optimal squared errors under Gaussian random designs over a wide range of sparsity classes. An appealing feature is that SLOPE does not require any knowledge of the degree of sparsity. This adaptivity to unknown sparsity has to do with the FDR control, which strikes the right balance between bias and variance. The proof of this result presents several elements not found in the high-dimensional statistics literature.\u003C\/p\u003E\u003Cp\u003E\u003Cstrong\u003EBio\u003C\/strong\u003E\u003C\/p\u003E\u003Cp\u003EWeijie Su is a fifth-year Ph.D. student in the \u003Ca href=\u0022https:\/\/statistics.stanford.edu\/\u0022\u003EStanford Statistics Department\u003C\/a\u003E, advised by \u003Ca href=\u0022http:\/\/statweb.stanford.edu\/~candes\/\u0022\u003EEmmanuel Cand\u00e8s\u003C\/a\u003E. In 2011, he received a B.S. in Mathematics and a B.A. in Economics (minor) from \u003Ca href=\u0022http:\/\/english.pku.edu.cn\/\u0022\u003EPeking University\u003C\/a\u003E. He spent three summers as an intern at \u003Ca href=\u0022http:\/\/research.microsoft.com\/en-us\/\u0022\u003EMicrosoft Research \u003C\/a\u003E(Beijing, 2010; Redmond, 2013; and Silicon Valley, 2014).\u003C\/p\u003E\u003Cp\u003EWeijie is broadly interested in high-dimensional statistics, convex optimization, and applied probability. His main focus is to develop and analyze model selection procedures that address challenges arising from high dimensionality, multicollinearity, and various forms of constraints in modern data analysis.\u003C\/p\u003E","summary":null,"format":"limited_html"}],"field_subtitle":"","field_summary":"","field_summary_sentence":[{"value":"Seminar - Weijie Su"}],"uid":"27187","created_gmt":"2016-01-26 11:47:12","changed_gmt":"2017-04-13 21:16:56","author":"Anita Race","boilerplate_text":"","field_publication":"","field_article_url":"","field_event_time":{"event_time_start":"2016-02-04T10:00:00-05:00","event_time_end":"2016-02-04T10:00:00-05:00","event_time_end_last":"2016-02-04T10:00:00-05:00","gmt_time_start":"2016-02-04 15:00:00","gmt_time_end":"2016-02-04 15:00:00","gmt_time_end_last":"2016-02-04 15:00:00","rrule":null,"timezone":"America\/New_York"},"extras":[],"groups":[{"id":"1242","name":"School of Industrial and Systems Engineering (ISYE)"}],"categories":[],"keywords":[],"core_research_areas":[],"news_room_topics":[],"event_categories":[{"id":"1795","name":"Seminar\/Lecture\/Colloquium"}],"invited_audience":[{"id":"78751","name":"Undergraduate students"},{"id":"78761","name":"Faculty\/Staff"},{"id":"174045","name":"Graduate students"}],"affiliations":[],"classification":[],"areas_of_expertise":[],"news_and_recent_appearances":[],"phone":[],"contact":[],"email":[],"slides":[],"orientation":[],"userdata":""}}}