47 min listen
OpenAI’s Sora team thinks we’ve only seen the "GPT-1 of video models"
FromNo Priors: Artificial Intelligence | Technology | Startups
OpenAI’s Sora team thinks we’ve only seen the "GPT-1 of video models"
FromNo Priors: Artificial Intelligence | Technology | Startups
ratings:
Length:
31 minutes
Released:
Apr 25, 2024
Format:
Podcast episode
Description
AI-generated videos are not just leveled-up image generators. But rather, they could be a big step forward on the path to AGI. This week on No Priors, the team from Sora is here to discuss Open AI’s recently announced generative video model, which can take a text prompt and create realistic, visually coherent, high-definition clips that are up to a minute long.
Sora team leads, Aditya Ramesh, Tim Brooks, and Bill Peebles join Elad and Sarah to talk about developing Sora. The generative video model isn’t yet available for public use but the examples of its work are very impressive. However, they believe we’re still in the GPT-1 era of AI video models and are focused on a slow rollout to ensure the model is in the best place possible to offer value to the user and more importantly they’ve applied all the safety measures possible to avoid deep fakes and misinformation. They also discuss what they’re learning from implementing diffusion transformers, why they believe video generation is taking us one step closer to AGI, and why entertainment may not be the main use case for this tool in the future.
Show Links:
Bling Zoo video
Man eating a burger video
Tokyo Walk video
Sign up for new podcasts every week. Email feedback to show@no-priors.com
Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @_tim_brooks l @billpeeb l @model_mechanic
Show Notes:
(0:00) Sora team Introduction
(1:05) Simulating the world with Sora
(2:25) Building the most valuable consumer product
(5:50) Alternative use cases and simulation capabilities
(8:41) Diffusion transformers explanation
(10:15) Scaling laws for video
(13:08) Applying end-to-end deep learning to video
(15:30) Tuning the visual aesthetic of Sora
(17:08) The road to “desktop Pixar” for everyone
(20:12) Safety for visual models
(22:34) Limitations of Sora
(25:04) Learning from how Sora is learning
(29:32) The biggest misconceptions about video models
Sora team leads, Aditya Ramesh, Tim Brooks, and Bill Peebles join Elad and Sarah to talk about developing Sora. The generative video model isn’t yet available for public use but the examples of its work are very impressive. However, they believe we’re still in the GPT-1 era of AI video models and are focused on a slow rollout to ensure the model is in the best place possible to offer value to the user and more importantly they’ve applied all the safety measures possible to avoid deep fakes and misinformation. They also discuss what they’re learning from implementing diffusion transformers, why they believe video generation is taking us one step closer to AGI, and why entertainment may not be the main use case for this tool in the future.
Show Links:
Bling Zoo video
Man eating a burger video
Tokyo Walk video
Sign up for new podcasts every week. Email feedback to show@no-priors.com
Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @_tim_brooks l @billpeeb l @model_mechanic
Show Notes:
(0:00) Sora team Introduction
(1:05) Simulating the world with Sora
(2:25) Building the most valuable consumer product
(5:50) Alternative use cases and simulation capabilities
(8:41) Diffusion transformers explanation
(10:15) Scaling laws for video
(13:08) Applying end-to-end deep learning to video
(15:30) Tuning the visual aesthetic of Sora
(17:08) The road to “desktop Pixar” for everyone
(20:12) Safety for visual models
(22:34) Limitations of Sora
(25:04) Learning from how Sora is learning
(29:32) The biggest misconceptions about video models
Released:
Apr 25, 2024
Format:
Podcast episode
Titles in the series (64)
What is the role of academia in modern AI research? With Stanford Professor Dr. Percy Liang: When AI research is evolving at warp speed and takes significant capital and compute power, what is the role of academia? Dr. Percy Liang – Stanford computer science professor and director of the Stanford Center for Research on Foundation Models talks about training costs, distributed infrastructure, model evaluation, alignment, and societal impact. Sarah Guo and Elad Gil join Percy at his office to discuss the evolution of research in NLP, why AI developers should aim for superhuman levels of performance, the goals of the Center for Research on Foundation Models, and Together, a decentralized cloud for artificial intelligence. by No Priors: Artificial Intelligence | Technology | Startups