Expérience et formation
Publications
-
Foundations of Sequence-to-Sequence Modeling for Time Series
AISTATS
The availability of large amounts of time series data, paired with the performance of deep-learning algorithms on a broad class of problems, has recently led to significant interest in the use of sequence-to-sequence models for time series forecasting. We provide the first theoretical analysis of this time series forecasting framework. We include a comparison of sequence-to-sequence modeling to classical time series models, and as such our theory can serve as a quantitative guide for…
The availability of large amounts of time series data, paired with the performance of deep-learning algorithms on a broad class of problems, has recently led to significant interest in the use of sequence-to-sequence models for time series forecasting. We provide the first theoretical analysis of this time series forecasting framework. We include a comparison of sequence-to-sequence modeling to classical time series models, and as such our theory can serve as a quantitative guide for practitioners choosing between different modeling methodologies.
Autres auteurs -
-
Learning Determinantal Point Processes by Sampling Inferred Negatives
AISTATS
Determinantal Point Processes (DPPs) have attracted significant interest from the machine-learning community due to their ability to elegantly and tractably model the delicate balance between quality and diversity of sets. We consider learning DPPs from data, a key task for DPPs; for this task, we introduce a novel optimization problem, Contrastive Estimation (CE), which encodes information about “negative” samples into the basic learning model. CE is grounded in the successful use of negative…
Determinantal Point Processes (DPPs) have attracted significant interest from the machine-learning community due to their ability to elegantly and tractably model the delicate balance between quality and diversity of sets. We consider learning DPPs from data, a key task for DPPs; for this task, we introduce a novel optimization problem, Contrastive Estimation (CE), which encodes information about “negative” samples into the basic learning model. CE is grounded in the successful use of negative information in machine-vision and language modeling. Depending on the chosen negative distribution (which may be static or evolve during optimization), CE assumes two different forms, which we analyze theoretically and experimentally. We evaluate our new model on real-world datasets; on a challenging dataset, CE learning delivers a considerable improvement in predictive performance over a DPP learned without using contrastive information.
Autres auteursVoir la publication -
Exponentiated Strongly Rayleigh Distributions
NeurIPS
Strongly Rayleigh (SR) measures are discrete probability distributions over the subsets of a ground set. They enjoy strong negative dependence properties, as a result of which they assign higher probability to subsets of diverse elements. We introduce in this paper Exponentiated Strongly Rayleigh (ESR) measures, which sharpen (or smoothen) the negative dependence property of SR measures via a single parameter (the exponent) that can intuitively understood as an inverse temperature. We develop…
Strongly Rayleigh (SR) measures are discrete probability distributions over the subsets of a ground set. They enjoy strong negative dependence properties, as a result of which they assign higher probability to subsets of diverse elements. We introduce in this paper Exponentiated Strongly Rayleigh (ESR) measures, which sharpen (or smoothen) the negative dependence property of SR measures via a single parameter (the exponent) that can intuitively understood as an inverse temperature. We develop efficient MCMC procedures for approximate sampling from ESRs, and obtain explicit mixing time bounds for two concrete instances: exponentiated versions of Determinantal Point Processes and Dual Volume Sampling. We illustrate some of the potential of ESRs, by applying them to a few machine learning tasks; empirical results confirm that beyond their theoretical appeal, ESR-based models hold significant promise for these tasks.
Autres auteursVoir la publication -
Maximizing Induced Cardinality Under a Determinantal Point Process
NeurIPS
Determinantal point processes (DPPs) are well-suited to recommender systems where the goal is to generate collections of diverse, high-quality items. In the existing literature this is usually formulated as finding the mode of the DPP (the so-called MAP set). However, the MAP objective inherently assumes that the DPP models “optimal” recommendation sets, and yet obtaining such a DPP is nontrivial when there is no ready source of example optimal sets. In this paper we advocate an alternative…
Determinantal point processes (DPPs) are well-suited to recommender systems where the goal is to generate collections of diverse, high-quality items. In the existing literature this is usually formulated as finding the mode of the DPP (the so-called MAP set). However, the MAP objective inherently assumes that the DPP models “optimal” recommendation sets, and yet obtaining such a DPP is nontrivial when there is no ready source of example optimal sets. In this paper we advocate an alternative framework for applying DPPs to recommender systems. Our approach assumes that the DPP simply models user engagements with recommended items, which is more consistent with how DPPs for recommender systems are typically trained. With this assumption, we are able to formulate a metric that measures the expected number of items that a user will engage with. We formalize this optimization of this metric as the Maximum Induced Cardinality (MIC) problem. Although the MIC objective is not submodular, we show that it can be approximated by a submodular function, and that empirically it is well-optimized by a greedy algorithm.
Autres auteurs -
Prix et distinctions
-
Google PhD Fellow in Machine Learning
Google
-
EECS Graduate Alumni Fellowship
MIT EECS department
-
Médaille d'argent au SWERC (South Western Europe Regional Programming Contest)
ACM-ICPC
Concours de programmation et d'algorithmique par équipes de trois personnes, réunissant des équipes venant de France, Suisse, Portugal, Italie, Espagne, Allemagne et Autriche
Langues
-
English
Bilingue ou langue natale
-
French
Bilingue ou langue natale
-
German
Capacité professionnelle générale
-
Japanese
Capacité professionnelle générale
-
Chinese
Notions
Autres profils similaires
-
Jerome Thai, PhD
Se connecter -
Nicolas Papernot
Se connecter -
Felix Grezes
Machine Learning Specialist | Harvard-Smithsonian Center for Astrophysics
Se connecter -
Danielle Tchuinkou Kwadjo
Ph.D. Computer Engineering. Hardware Accelerator of DNN on FPGAs, C/C++| Verilog | LLVM | Python | bash | Tensorflow
Se connecter -
Fujun Luan
Se connecter -
Mehdi Boukhechba, PhD
Sr Principal Scientist | Data Science & Digital Health | J&J Innovative Medicine
Se connecter -
Jonathan Z Amar
GenAI @ Verily (ex Google Life Sciences) | PhD
Se connecter -
Francois Hamon, PhD
Se connecter -
Alexandre Barachant
Se connecter -
Bertrand Delorme
Se connecter