Jump to Content
Kishore Papineni

Kishore Papineni

Kishore leads the Coauthor team whose objective is cross-lingual cross-modal access to dynamically organized information. His team hopes to make content consumption or creation a richer experience by surfacing relevant and diverse information from the web, possibly synthesized dynamically from across different sources or types of content such as text, images, charts, and videos. Coauthor team powers the web content suggestions in Google Docs when users are writing a document, and is working on additional content recommendation applications. His work at Google includes veracity of information on the web, depth of discourse on a topic in a document, drift of discourse on a topic on the web, identifying concepts peculiar to a collection of documents and relationships among the concepts, and identifying different perspectives in content. His past work was in the areas of automatic control theory, natural language understanding, dialog management, machine translation, and display advertisements. Prior to joining Google, he led machine learning at Yahoo! Research and machine translation at IBM Research. He is a coauthor of the BLEU metric for automatic evaluation of machine translation quality (awarded 2018 Test-of-Time Paper on Computational Linguistics). He was a founding Editor-in-Chief of ACM Transactions on Speech and Language Processing from 2003-2007.
Authored Publications
Google Publications
Other Publications
Sort By
  • Title
  • Title, descending
  • Year
  • Year, descending
    Preview abstract It is generally believed that robust training of extremely large networks is critical to their success in real-world applications. However, when taken to the extreme, methods that promote robustness can hurt the model's sensitivity to rare or underrepresented patterns. In this paper, we discuss this trade-off between sensitivity and robustness to natural (non-adversarial) perturbations by introducing two notions: contextual feature utility and contextual feature sensitivity. We propose Feature Contrastive Learning (FCL) that encourages a model to be more sensitive to the features that have higher contextual utility. Empirical results demonstrate that models trained with FCL achieve a better balance of robustness and sensitivity, leading to improved generalization in the presence of noise on both vision and NLP datasets. View details
    Preview abstract Document and discourse segmentation are two fundamental NLP tasks pertaining to breaking up text into constituents, which are commonly used to help downstream tasks such as information retrieval or text summarization. In this work, we propose three transformer-based architectures and provide comprehensive comparisons with previously proposed approaches on three standard datasets. We establish a new state-of-the-art, reducing in particular the error rates by a large margin in all cases. We further analyze model sizes and find that we can build models with many fewer parameters while keeping good performance, thus facilitating real-world applications. View details
    Maximally representative allocations for guaranteed delivery advertising campaigns
    R. Preston McAfee
    Review of Economic Design, vol. 17 (2013), pp. 83-94
    Bidding for Representative Allocations for Display Advertising
    Arpita Ghosh
    Randolph Preston McAfee
    WINE (2009), pp. 208-219
    Bidding for Representative Allocations for Display Advertising
    Arpita Ghosh
    Randolph Preston McAfee
    CoRR, vol. abs/0910.0880 (2009)