- Film & Animation
- Music
- Pets & Animals
- Sports
- Travel & Events
- Gaming
- People & Blogs
- Comedy
- Entertainment
- News & Politics
- How-to & Style
- Non-profits & Activism
- Lifestyle
- Krew
- Learning
- Teaching
- Robotics
- Drones
- Ai
- Crypto
- DAW
- Live
- Podcasts
- Webinars
- Courses
- Couching
- Teachers
- Schools
- Tutorials
- Tutoring
- Class
- Watch Party
- Whiteboard
- Meeting
- Live
- Livestream
- Automation
- Robotics
- AR/VR
- Artificial Intelligence
- Open Source
- Developer
- Consultant
- Entrepreneur
- Meeting Room
- Rights Managed (RM)
- Royalty Free License (RF)
- Editorial Use License
- Creative Commons License
- Public Domain
- Sales
- Donation
- Generate Ai Image
- Generate Ai Post
- Generative AI
- Generate Ai Speech
- Generate Ai Music
- Generate Ai Video
- Albums
- Articles
- Marketplace
- Vlogging
- Students
- Authors
- Journalist
- Virtual Background
- Free Content
- Google Analytics
- Virtual Whiteboard
- Ai Noise Cancellation
- Celebrities
- Influencers
- Movies
- Aeronautics
- Aviation
- Science
- Space Agency
- Machine Learning
- Science
- Space
- Satellite
- Humanoids
- Aliens
- Android
- Workshops
- Future
- Graphic Design
- Music Production
- Entertainment Business
- Screenwriting
- Production
- Editing
- Audiography
- Filmmaking Courses
- Academy Award
- Emmy
- Documentaries
- TV commercials
- Independent Film
- Filmmakers
- On-Set
- DIY
- Television
- Academy
- Film School
- Instructors
- Training
- Books
- Certification
- Film Educator
- Pictures
- Stidio
- 4K
- Video Maker
- Hollywood
- Special Effects
- Learners
- Online Courses
- Virtual Classrooms
- Classroom
- Coding
- Photography
- Tutorials
- Portraits
- Black & White
- Directing
- Director
- Degree Programs
- Classes
- Online Movies
- Game Design
- Acting
- Script
- Script Writing
- MasterClass
- Story
- Storyboard
- Stage
- Crew
- Casting
- Props
- Advertisement
- Grand Openings
- Ebook
- Events
- Premier
- Trailers
- Playlist
- Downloads
- Clips
- Rhythm & Blues
- World Beats
- Downtempo
- Train
- Stocks
- Cinematography
- Online Workshops
- Sound Design
- Film Scoring
- Visual Arts
- Film Production
- AI Filmmaking
- Guerilla Filmmaking
- Camera
- Shooting
- Widescreen
- Drawing
- Language
- Academics
- Private Lessons
- Physics
- Math
- Writing
- Coaching App
- Nutrition
- Personal Training
- Development
- Development
- Training Course Templates
- Synthesis
- Investment
- Finance
- eWallet
- CoinPayments
- Coinbase
- Alipay
- M-PESA
- Payments
- Payment Gateways
- Money Remittances
- Digital Payments
- Other
Lesson 9: Deep Learning Foundations to Stable Diffusion, 2022
(All lesson resources are available at http://course.fast.ai.) This is the first lesson of part 2 of Practical Deep Learning for Coders. It starts with a tutorial on how to use pipelines in the Diffusers library to generate images. Diffusers is (in our opinion!) the best library available at the moment for image generation. It has many features and is very flexible. We explain how to use its many features, and discuss options for accessing the GPU resources needed to use the library.
We talk about some of the nifty tweaks available when using Stable Diffusion in Diffusers, and show how to use them: guidance scale (for varying the amount the prompt is used), negative prompts (for removing concepts from an image), image initialisation (for starting with an existing image), textual inversion (for adding your own concepts to generated images), Dreambooth (an alternative approach to textual inversion).
The second half of the lesson covers the key concepts involved in Stable Diffusion:
- CLIP embeddings
- The VAE (variational autoencoder)
- Predicting noise with the unet
- Removing noise with schedulers.
You can discuss this lesson, and access links to all notebooks and resources from it, at this forum topic: https://forums.fast.ai/t/lesso....n-9-part-2-preview/1
0:00 - Introduction
6:38 - This course vs DALL-E 2
10:38 - How to take full advantage of this course
12:14 - Cloud computing options
14:58 - Getting started (Github, notebooks to play with, resources)
20:48 - Diffusion notebook from Hugging Face
26:59 - How stable diffusion works
30:06 - Diffusion notebook (guidance scale, negative prompts, init image, textual inversion, Dreambooth)
45:00 - Stable diffusion explained
53:04 - Math notation correction
1:14:37 - Creating a neural network to predict noise in an image
1:27:46 - Working with images and compressing the data with autoencoders
1:40:12 - Explaining latents that will be input into the unet
1:43:54 - Adding text as one hot encoded input to the noise and drawing (aka guidance)
1:47:06 - How to represent numbers vs text embeddings in our model with CLIP encoders
1:53:13 - CLIP encoder loss function
2:00:55 - Caveat regarding "time steps"
2:07:04 Why don’t we do this all in one step?
Thanks to fmussari for the transcript, and to Raymond-Wu (on forums.fast.ai) for the timestamps.