11 Folgen

L'apprentissage statistique joue de nos jours un rôle croissant dans de nombreux domaines scientifiques et doit de ce fait faire face à des problèmes nouveaux. Il est par conséquent important de proposer des méthodes d'apprentissage statistique adaptées aux problèmes modernes posés par les différents champs d'application. Outre l'importance de la précision des méthodes proposées, elles devront également apporter une meilleure compréhension des phénomènes observés. Afin de faciliter les contacts entre les différentes communautés et de faire ainsi germer de nouvelles idées, un colloquium d'audience internationale (en langue anglaise) sur le thème «Challenging problems in Statistical Learning» a été organisé à l'Université Paris 1 les 5 et 6 avril 2012. Vous trouverez ci-dessous les enregistrements des exposés donnés lors de ce colloquium.
Ce colloquium a été organisé par C. Bouveyron, Christophe Biernacki , Alain Célisse , Serge Iovle & Julien Jacques (Laboratoire SAMM, Paris 1, Laboratoire Paul Painlevé, Université Lille 1, CNRS & Modal, INRIA), avec le soutien de la SFdS.
Recommandé à : étudiant de la discipline, chercheur - Catégorie : cours podcast - Année de réalisation : 2012

StatLearn 2012 - Workshop on "Challenging problems in Statistical Learning‪"‬ Statlearn2012

    • Bildung

L'apprentissage statistique joue de nos jours un rôle croissant dans de nombreux domaines scientifiques et doit de ce fait faire face à des problèmes nouveaux. Il est par conséquent important de proposer des méthodes d'apprentissage statistique adaptées aux problèmes modernes posés par les différents champs d'application. Outre l'importance de la précision des méthodes proposées, elles devront également apporter une meilleure compréhension des phénomènes observés. Afin de faciliter les contacts entre les différentes communautés et de faire ainsi germer de nouvelles idées, un colloquium d'audience internationale (en langue anglaise) sur le thème «Challenging problems in Statistical Learning» a été organisé à l'Université Paris 1 les 5 et 6 avril 2012. Vous trouverez ci-dessous les enregistrements des exposés donnés lors de ce colloquium.
Ce colloquium a été organisé par C. Bouveyron, Christophe Biernacki , Alain Célisse , Serge Iovle & Julien Jacques (Laboratoire SAMM, Paris 1, Laboratoire Paul Painlevé, Université Lille 1, CNRS & Modal, INRIA), avec le soutien de la SFdS.
Recommandé à : étudiant de la discipline, chercheur - Catégorie : cours podcast - Année de réalisation : 2012

    • video
    1.1 Dimension reduction based on finite mixture modeling of inverse regression (Luca Scrucca)

    1.1 Dimension reduction based on finite mixture modeling of inverse regression (Luca Scrucca)

    Consider the usual regression problem in which we want to study the conditional distribution of a response Y given a set of predictors X. Sufficient dimension reduction (SDR) methods aim at replacing the high-dimensional vector of predictors by a lower-dimensional function R(X) with no loss of information about the dependence of the response variable on the predictors. Almost all SDR methods restrict attention to the class of linear reductions, which can be represented in terms of the projection of X onto a dimension-reduction subspace (DRS). Several methods have been proposed to estimate the basis of the DRS, such as sliced inverse regression (SIR; Li, 1991), principal Hessian directions (PHD; Li, 1992), sliced average variance estimation (SAVE; Cook and Weisberg, 1991), directional regression (DR; Li et al., 2005) and inverse regression estimation (IRE; Cook and Ni, 2005). A novel SDR method, called MSIR, based on finite mixtures of Gaussians has been recently proposed (Scrucca, 2011) as an extension to SIR. The talk will present the MSIR methodology and some recent advances. In particular, a BIC criterion for the selection the dimensionality of DRS will be introduced, and its extension for the purpose of variable selection. Finally, the application of MSIR in classification problems, both supervised and semi-supervised, will be discussed.

    • 1 Std.
    • video
    1.2 Information Visualization: An Introduction to the Field and Applications for Statistics (Petra Isenberg)

    1.2 Information Visualization: An Introduction to the Field and Applications for Statistics (Petra Isenberg)

    Information visualization is a research area that focuses on making structures and content of large and complex data sets visually understandable and interactively analyzable. The goal of information visualization tools and techniques is to increase our ability to gain insight and make decisions for many types of datasets, tasks, and analysis scenarios. With the increase in size and complexity of data sets today, the research area of information visualization increasingly gains in importance and recognition. In this talk I will present principles of data representation and interaction and introduce a number of existing applications, tools, and techniques and show how they can be applied to questions in statistics and statistical learning.

    • 46 Min.
    • video
    2.1 Hypothesis Testing and Bayesian Inference: New Applications of Kernel Methods (Arthur Gretton)

    2.1 Hypothesis Testing and Bayesian Inference: New Applications of Kernel Methods (Arthur Gretton)

    In the early days of kernel machines research, the "kernel trick" was considered a useful way of constructing nonlinear learning algorithms from linear ones, by applying the linear algorithms to feature space mappings of the original data. Recently, it has become clear that a potentially more far reaching use of kernels is as a linear way of dealing with higher order statistics, by mapping probabilities to a suitable reproducing kernel Hilbert space (i.e., the feature space is an RKHS). I will describe how probabilities can be mapped to reproducing kernel Hilbert spaces, and how to compute distances between these mappings. A measure of strength of dependence between two random variables follows naturally from this distance. Applications that make use of kernel probability embeddings include: - Nonparametric two-sample testing and independence testing in complex (high dimensional) domains. As an application, we find whether text in English is translated from the French, as opposed to being random extracts on the same topic. - Bayesian inference, in which the prior and likelihood are represented as feature space mappings, and a posterior feature space mapping is obtained. In this case, Bayesian inference can be undertaken even in the absence of a model, by learning the prior and likelihood mappings from samples.

    • 57 Min.
    • video
    2.2 Functional estimation in high dimensional data : Application to classification (Sophie Dabo-Niang)

    2.2 Functional estimation in high dimensional data : Application to classification (Sophie Dabo-Niang)

    Functional data are becoming increasingly common in a variety of fields. Many studies underline the importance to consider the representation of data as functions. This has sparked a growing attention in the development of adapted statistical tools that allow to analyze such kind of data : functional data analysis (FDA). The aims of FDA are mainly the same as in the classical statistical analysis, e.g. representing and visualizing the data, studying variability and trends, comparing different data sets, as well as modeling and predicting,... Recent advances in FDA allow to construct different classification methods, based on the comparison between centrality curves or using change points,... We review some procedures that have been used to classify functional data. The main point is to show the good practical behaviors of these procedures on a sample of curves. In addition, theoretical advances on functional estimations related to these classification methods are provided.

    • 46 Min.
    • video
    2.3 Discriminative clustering for high-dimensional data (Camille Brunet)

    2.3 Discriminative clustering for high-dimensional data (Camille Brunet)

    A new family of 12 probabilistic models, introduced recently, aims to simultaneously cluster and visualize high-dimensional data. It is based on a mixture model which fits the data into a latent discriminative subspace with an intrinsic dimension bounded by the number of clusters. An estimation procedure, named the Fisher-EM algorithm has also been proposed and turns out to outperform other subspace clustering in most situations. Moreover the convergence properties of the Fisher-EM algorithm are discussed; in particular it is proved that the algorithm is a GEM algorithm and converges under weak conditions in the general case. Finally, a sparse extension of the Fisher-EM algorithm is proposed in order to perform a selection of the original variables which are discriminative.

    • 56 Min.
    • video
    3.1 Exploring Clustering Structure in Ranking Data (Brendan Murphy)

    3.1 Exploring Clustering Structure in Ranking Data (Brendan Murphy)

    Cluster analysis is concerned with finding homogeneous groups in a population. Model-based clustering methods provide a framework for developing clustering methods through the use of statistical models. This approach allows for uncertainty to be quantified using probability and for the properties of a clustering method to be understood on the basis of a well defined statistical model. Mixture models provide a basis for many model-based clustering methods. Ranking data arise when judges rank some or all of a set of objects. Examples of ranking data include voting data from elections that use preferential voting systems (eg. PR-STV) and customer preferences for products in marketing applications. A mixture of experts model is a mixture model in which the model parameters are functions of covariates. We explore the use of mixture of experts models in cluster analysis, so that clustering can be better understood. The choice of how and where covariates enter the mixture of experts model has implications for the clustering performance and the interpretation of the results. The use of covariates in clustering is demonstrated on examples from studying voting blocs in elections and examining customer segments marketing.

    • 1 Std. 4 Min.

Top‑Podcasts in Bildung

Eine Stunde History - Deutschlandfunk Nova
Deutschlandfunk Nova
Gehirn gehört - Prof. Dr. Volker Busch
Prof. Dr. Volker Busch
G Spot - mit Stefanie Giesinger
Stefanie Giesinger & Studio Bummens
Easy German: Learn German with native speakers | Deutsch lernen mit Muttersprachlern
Cari, Manuel und das Team von Easy German
ZEIT Sprachen – English, please!
ZEIT ONLINE
Quarks Science Cops
Quarks

Mehr von Université Paris 1 Panthéon-Sorbonne

Découper le temps : les périodes de l'histoire
Université de Paris 1 Panthéon-Sorbonne
L'industrie, patrimoine et culture (2010-2011)
Université Paris 1 Panthéon-Sorbonne
Droit des entreprises
Bruno Dondero
Droit constitutionnel et institutions politiques (CAVEJ, Michel Verpaux, 2010)
Université Paris 1 Panthéon-Sorbonne
Témoignages EPI
Université Paris 1 Panthéon-Sorbonne
Biodiversité
UVED