Loading...
background pattern
Microsoft Research Blog

Research Focus: Week of October 28, 2024 

November 1, 2024

New Research | FLASH: Workflow automation agent for diagnosing recurring incidents; METAREFLECTION: Learning instructions for language agents using past reflections; Boosting LLM training efficiency through faster communication between GPUs; and more.

"2023 Microsoft Research Year In Review" in white text on a blue, green, and purple abstract gradient background
Microsoft Research Blog

Research at Microsoft 2023: A year of groundbreaking AI advances and discoveries 

December 22, 2023

AI saw unparalleled growth in 2023, reaching millions daily. This progress owes much to the extensive work of Microsoft researchers and collaborators. In this review, learn about the advances in 2023, which set the stage for further progress in 2024.

In the news | Metaverse Post

Microsoft’s DeepSpeed4Science Advances AI in Scientific Research 

September 22, 2023

Microsoft recently launched DeepSpeed4Science initiative to apply deep learning in natural sciences, including drug development and renewable energy. DeepSpeed, an open-source AI framework, aims to accelerate and scale up deep learning processes.

In the news | WinBuzzer

Microsoft Announces DeepSpeed4Science Initiative for AI Based Scientific Research 

September 20, 2023

Microsoft has introduced the DeepSpeed4Science initiative through its DeepSpeed team. The initiative focuses on the application of deep learning in the natural sciences, targeting areas such as drug development and renewable energy. The DeepSpeed system, an open-source AI framework from…

DeepSpeed4Science Initiative - graphic with 6 icons
Microsoft Research Blog

Announcing the DeepSpeed4Science Initiative: Enabling large-scale scientific discovery through sophisticated AI system technologies 

September 19, 2023 | Shuaiwen Leon Song, Bonnie Kruft, Minjia Zhang, Conglong Li, Martin Cai, and Yuxiong He

Editor’s note, Sept. 28, 2023 – The founding collaborators list was updated to correct omissions and the scientific foundation model graph was updated to correct information. In the next decade, deep learning may revolutionize the natural sciences, enhancing our capacity to…

DeepSpeed ZeRO++ blog hero
Microsoft Research Blog

DeepSpeed ZeRO++: A leap in speed for LLM and chat model training with 4X less communication 

June 22, 2023 | DeepSpeed Team and Andrey Proskurin

Large AI models are transforming the digital world. Generative language models like Turing-NLG, ChatGPT, and GPT-4, powered by large language models (LLMs), are incredibly versatile, capable of performing tasks like summarization, coding, and translation. Similarly, large multimodal generative models like…

Microsoft Research Focus 03: Week of November 7th, 2022
Microsoft Research Blog

Research Focus: Week of November 7, 2022 

November 8, 2022

Welcome to Research Focus, a new series of blog posts that highlights notable publications, events, code/datasets, new hires and other milestones from across the research community at Microsoft. Barun Patra, Saksham Singhal, Shaohan Huang, Zewen Chi, Li Dong, Furu Wei,…

Three bar plots. The first plot shows that the model size of XTC-BERT is 32 times smaller than that of BERT, and two dots show the accuracy of BERT and XTC-BERT, which are 83.95 and 83.44, respectively. The second one shows that INT8 using ZeroQuant can be 2.6 times faster than Baseline with FP16 using PyTorch and ZeoQuant can reduce the number of GPUs for inference from 2 to 1, which in total provides 5.2 times efficiency. It also shows that ZeroQuant has 50.4 accuracy compared to 50.5 using Baseline PyTorch. The third plot shows that ZeroQuant is more than 5000 times cheaper than baseline to compress a model, and the accuracy of ZeroQuant is 42.26 compared to 42.35 of baseline.
Microsoft Research Blog

DeepSpeed Compression: A composable library for extreme compression and zero-cost quantization 

July 20, 2022 | DeepSpeed Team and Andrey Proskurin

Large-scale models are revolutionizing deep learning and AI research, driving major improvements in language understanding, generating creative texts, multi-lingual translation and many more. But despite their remarkable capabilities, the models’ large size creates latency and cost constraints that hinder the…

DeepSpeed shares findings and innovations for MoE models and systems that 1) reduce training cost by 5x, 2) reduce MoE parameter size by up to 3.7x and 3) reduce MoE inference latency by 7.3x at an unprecedented scale and offer up to 4.5x faster and 9x cheaper inference for MoE models compared to quality-equivalent dense models.
Microsoft Research Blog

DeepSpeed: Advancing MoE inference and training to power next-generation AI scale 

January 19, 2022 | DeepSpeed Team and Andrey Proskurin

In the last three years, the largest trained dense models have increased in size by over 1,000 times, from a few hundred million parameters to over 500 billion parameters in Megatron-Turing NLG 530B (MT-NLG). Improvements in model quality with size…

  • Previous
  • 1
  • 2
  • 3
  • 4
  • Next