France
1 k abonnés + de 500 relations

Devenir membre pour voir le profil

Expérience et formation

  • Bioptimus

Voir toute l’expérience de Zelda

Découvrez son poste, son ancienneté et plus encore.

ou

En cliquant sur Continuer pour vous inscrire ou vous identifier, vous acceptez les Conditions d’utilisation, la Politique de confidentialité et la Politique relative aux cookies de LinkedIn.

Publications

  • Foundations of Sequence-to-Sequence Modeling for Time Series

    AISTATS

    The availability of large amounts of time series data, paired with the performance of deep-learning algorithms on a broad class of problems, has recently led to significant interest in the use of sequence-to-sequence models for time series forecasting. We provide the first theoretical analysis of this time series forecasting framework. We include a comparison of sequence-to-sequence modeling to classical time series models, and as such our theory can serve as a quantitative guide for…

    The availability of large amounts of time series data, paired with the performance of deep-learning algorithms on a broad class of problems, has recently led to significant interest in the use of sequence-to-sequence models for time series forecasting. We provide the first theoretical analysis of this time series forecasting framework. We include a comparison of sequence-to-sequence modeling to classical time series models, and as such our theory can serve as a quantitative guide for practitioners choosing between different modeling methodologies.

    Autres auteurs
    • Vitaly Kuznetsov
    Voir la publication
  • Learning Determinantal Point Processes by Sampling Inferred Negatives

    AISTATS

    Determinantal Point Processes (DPPs) have attracted significant interest from the machine-learning community due to their ability to elegantly and tractably model the delicate balance between quality and diversity of sets. We consider learning DPPs from data, a key task for DPPs; for this task, we introduce a novel optimization problem, Contrastive Estimation (CE), which encodes information about “negative” samples into the basic learning model. CE is grounded in the successful use of negative…

    Determinantal Point Processes (DPPs) have attracted significant interest from the machine-learning community due to their ability to elegantly and tractably model the delicate balance between quality and diversity of sets. We consider learning DPPs from data, a key task for DPPs; for this task, we introduce a novel optimization problem, Contrastive Estimation (CE), which encodes information about “negative” samples into the basic learning model. CE is grounded in the successful use of negative information in machine-vision and language modeling. Depending on the chosen negative distribution (which may be static or evolve during optimization), CE assumes two different forms, which we analyze theoretically and experimentally. We evaluate our new model on real-world datasets; on a challenging dataset, CE learning delivers a considerable improvement in predictive performance over a DPP learned without using contrastive information.

    Autres auteurs
    Voir la publication
  • Exponentiated Strongly Rayleigh Distributions

    NeurIPS

    Strongly Rayleigh (SR) measures are discrete probability distributions over the subsets of a ground set. They enjoy strong negative dependence properties, as a result of which they assign higher probability to subsets of diverse elements. We introduce in this paper Exponentiated Strongly Rayleigh (ESR) measures, which sharpen (or smoothen) the negative dependence property of SR measures via a single parameter (the exponent) that can intuitively understood as an inverse temperature. We develop…

    Strongly Rayleigh (SR) measures are discrete probability distributions over the subsets of a ground set. They enjoy strong negative dependence properties, as a result of which they assign higher probability to subsets of diverse elements. We introduce in this paper Exponentiated Strongly Rayleigh (ESR) measures, which sharpen (or smoothen) the negative dependence property of SR measures via a single parameter (the exponent) that can intuitively understood as an inverse temperature. We develop efficient MCMC procedures for approximate sampling from ESRs, and obtain explicit mixing time bounds for two concrete instances: exponentiated versions of Determinantal Point Processes and Dual Volume Sampling. We illustrate some of the potential of ESRs, by applying them to a few machine learning tasks; empirical results confirm that beyond their theoretical appeal, ESR-based models hold significant promise for these tasks.

    Autres auteurs
    Voir la publication
  • Maximizing Induced Cardinality Under a Determinantal Point Process

    NeurIPS

    Determinantal point processes (DPPs) are well-suited to recommender systems where the goal is to generate collections of diverse, high-quality items. In the existing literature this is usually formulated as finding the mode of the DPP (the so-called MAP set). However, the MAP objective inherently assumes that the DPP models “optimal” recommendation sets, and yet obtaining such a DPP is nontrivial when there is no ready source of example optimal sets. In this paper we advocate an alternative…

    Determinantal point processes (DPPs) are well-suited to recommender systems where the goal is to generate collections of diverse, high-quality items. In the existing literature this is usually formulated as finding the mode of the DPP (the so-called MAP set). However, the MAP objective inherently assumes that the DPP models “optimal” recommendation sets, and yet obtaining such a DPP is nontrivial when there is no ready source of example optimal sets. In this paper we advocate an alternative framework for applying DPPs to recommender systems. Our approach assumes that the DPP simply models user engagements with recommended items, which is more consistent with how DPPs for recommender systems are typically trained. With this assumption, we are able to formulate a metric that measures the expected number of items that a user will engage with. We formalize this optimization of this metric as the Maximum Induced Cardinality (MIC) problem. Although the MIC objective is not submodular, we show that it can be approximated by a submodular function, and that empirically it is well-optimized by a greedy algorithm.

    Autres auteurs
    • Alex Kulesza
    • Jennifer Gillenwater
    • Sergei Vassilvitskii
    Voir la publication
  • Elementary Symmetric Polynomials for Optimal Experimental Design

    Neural Information Processing Systems

    Autres auteurs
  • Fixed-point algorithms for learning determinantal point processes

    ICML 2015

    Autres auteurs

Prix et distinctions

  • Google PhD Fellow in Machine Learning

    Google

  • EECS Graduate Alumni Fellowship

    MIT EECS department

  • Médaille d'argent au SWERC (South Western Europe Regional Programming Contest)

    ACM-ICPC

    Concours de programmation et d'algorithmique par équipes de trois personnes, réunissant des équipes venant de France, Suisse, Portugal, Italie, Espagne, Allemagne et Autriche

Langues

  • English

    Bilingue ou langue natale

  • French

    Bilingue ou langue natale

  • German

    Capacité professionnelle générale

  • Japanese

    Capacité professionnelle générale

  • Chinese

    Notions

Voir le profil complet de Zelda

  • Découvrir vos relations en commun
  • Être mis en relation
  • Contacter Zelda directement
Devenir membre pour voir le profil complet

Autres profils similaires