Image Generation/Synthesis
In the recent surge of AI applications, the image generation field has risen in popularity alongside large language models. Stable Diffusion has played a crucial role in this development.
Learning Resources:
OverView of generative AI animation techniques
How Stable Diffusion Works from Chris McCormick
How Diffusion Models Work from DeepLearning
AI Data Poisoning Tool - Nightshade
Nightshade is an innovative tool developed by researchers at the University of Chicago to protect artists' work from being used without permission in AI training datasets. It works by adding subtle modifications to images that are imperceptible to humans but can disrupt AI models that attempt to learn from or replicate these images. When AI systems train on Nightshade-protected images, they may produce distorted or incorrect outputs, effectively "poisoning" the model and deterring unauthorized use of artists' creations.
Last updated
Was this helpful?