## Dreamfusion
- text-to-3D model
- uses a pretrained 2D text-to-image difusion model to perform textto-3D synthesis
- Dreamfusion replaces previous CLIP techniques with a loss derived from distillation of a 2D difusion model
- the difusion model can be used as a loss within a generic continuous optimization problem to generate samples
- sampling in parameter space is much harder than in pixels as we want to create 3D models that look like good images when rendered from random angles
- this model uses a diferentiable generator - Other approaches are focused on sampling pixels, however, this model instead focuses on creating 3D models that look like good images when rendered from random angles