Microsoft Research Blog

English

  1. AID landing page banner

    AI for Domains (AID) 

    July 12, 2024

    AI for Domains (AID) Our mission The M365 Research organization's AI for Domains (AID) team is an applied research group focused on leveraging cutting-edge research to extend the capabilities, efficiency, reliability, and reduce risks associated with Copilot. We prioritize user privacy and confidentiality while unlocking…

  2. Advances in Natural Language Generation for Indian Languages 

    July 12, 2024

    Much of recent progress for natural language generation (NLG) has been in the context of English and, in general, high resource languages, however, Indian languages have yet to see similar paradigm shifts despite their speaking population comprising about a fifth of the world's population. Two…

  3. CodePlan: Repository-level Coding using LLMs and Planning 

    July 12, 2024

    Software engineering activities such as package migration, fixing error reports from static analysis or testing, and adding type annotations or other specifications to a codebase, involve pervasively editing the entire repository of code. We formulate these activities as repository-level coding tasks. Recent tools like GitHub…

  4. Intelligence Toolkit 

    July 11, 2024 | Dayenne Souza and Darren Edge

    The Intelligence Toolkit is a suite of interactive workflows for creating AI intelligence reports from real-world data sources. The toolkit is designed to help users identify patterns, answers, relationships, and risks within complex datasets, with generative AI (OpenAI GPT models) used to create reports on…

  5. Autoregressive Speech Synthesis without Vector Quantization 

    July 11, 2024

    We present MELLE, a novel continuous-valued tokens based language modeling approach for text to speech synthesis (TTS). MELLE autoregressively generates continuous mel-spectrogram frames directly from text condition, bypassing the need for vector quantization, which are originally designed for audio compression and sacrifice fidelity compared to…

  6. Accuracy is Not All You Need 

    July 11, 2024 | Abhinav Dutta, Sanjeev Krishnan, Nipun Kwatra, and Ramachandran Ramjee

    When Large Language Models (LLMs) are compressed using techniques such as quantization, the predominant way to demonstrate the validity of such techniques is by measuring the model's accuracy on various benchmarks.If the accuracies of the baseline model and the compressed model are close, it is…