AI Frontiers: The future of scale with Ahmed Awadallah and Ashley Llorens
Listen now
Description
Powerful large-scale AI models like GPT-4 are showing dramatic improvements in reasoning, problem-solving, and language capabilities. This marks a phase change for artificial intelligence—and a signal of accelerating progress to come.  In this Microsoft Research Podcast series, AI scientist and engineer Ashley Llorens (https://www.microsoft.com/en-us/research/people/allorens/) hosts conversations with his collaborators and colleagues about what these models—and the models that will come next—mean for our approach to creating, understanding, and deploying AI, its applications in areas such as healthcare and education, and its potential to benefit humanity.This episode features Senior Principal Research Manager Ahmed H. Awadallah (https://www.microsoft.com/en-us/research/people/hassanam/), whose work improving the efficiency of large-scale AI models and efforts to help move advancements in the space from research to practice have put him at the forefront of this new era of AI. Awadallah discusses the shift in dynamics between model size and amount—and quality—of data when it comes to model training; the recently published paper “Orca: Progressive Learning from Complex Explanation Traces of GPT-4,” which further explores the use of large-scale AI models to improve the performance of smaller, less powerful ones; and the need for better evaluation strategies, particularly as we move into a future in which Awadallah hopes to see gains in these models’ ability to continually learn.Learn more:* Orca: Progressive Learning from Complex Explanation Traces of GPT-4 (https://www.microsoft.com/en-us/research/publication/orca-progressive-learning-from-complex-explanation-traces-of-gpt-4/), June 2023 * Textbooks Are All You Need II: phi-1.5 technical report (https://www.microsoft.com/en-us/research/publication/textbooks-are-all-you-need-ii-phi-1-5-technical-report/), September 2023* AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation Framework (https://www.microsoft.com/en-us/research/publication/autogen-enabling-next-gen-llm-applications-via-multi-agent-conversation-framework/), August 2023 * LIDA: Automatic Generation of Grammar-Agnostic Visualizations and Infographics using Large Language Models (https://www.microsoft.com/en-us/research/publication/lida-automatic-generation-of-grammar-agnostic-visualizations-and-infographics-using-large-language-models/), March 2023* AI Explainer: Foundation models ​and the next era of AI (https://www.microsoft.com/en-us/research/blog/ai-explainer-foundation-models-and-the-next-era-of-ai/), March 2023 * AI and Microsoft Research (https://www.microsoft.com/en-us/research/focus-area/ai-and-microsoft-research/)
More Episodes
Research manager Karin Strauss and members of the DNA Data Storage Project reflect on the path to developing a synthetic DNA–based system for archival data storage, including the recent open-source release of its most powerful algorithm for DNA error correction.Get the Trellis BMA code: GitHub -...
Published 11/19/24
Published 11/19/24
The efficient simulation of molecules has the potential to change how the world understands biological systems and designs new drugs and biomaterials. Tong Wang discusses AI2BMD, an AI-based system designed to simulate large biomolecules with speed and accuracy.Read the paperGet the code
Published 11/14/24