Please read http://madebyoll.in/posts/world_emulation_via_neural_network.
You can also find more experimental neural worlds at https://neuralworlds.net.
Please read http://madebyoll.in/posts/world_emulation_via_neural_network.
You can also find more experimental neural worlds at https://neuralworlds.net.
jupyterlab_vim adds a truly unfortunate keybinding where, if you are in normal mode and press -, the current cell splits into two.
I have never intentionally activated this feature yet have unintentionally activated it many dozens of times. It's awful.
Unfortunately, the logical route to disable the split-cell feature (modifying keybinding config files) doesn't work. I don't know why.
Therefore, I now manually locate the vim_bindings file and patch it:
#!/usr/bin/env python3
from pathlib import PathGuided diffusion sampling typically uses two forward passes per step:
E[flow | noisy image, noise level, caption]E[flow | noisy image, noise level]These results are then linearly combined to form a single guided/superconditioned flow prediction.
These useful concepts show up in specific areas of NN-training literature but can be applied pretty broadly.
torch.rand(batch_size), you can use th.randperm(batch_size).add_(th.rand(batch_size)).div_(batch_size) instead, which has the same distribution but lower variance, and therefore trains more stably. This shows up in k-diffusion https://github.com/crowsonkb/k-diffusion/commit/a2b7b5f1ea0d3711a06661ca9e41b4e6089e5707, but it's applicable whenever you're randomizing data across the batch axis.TL;DR - I think the paper is a good contribution and basically holds up, but Figure 2 seems suspicious and the released repo doesn't include the pieces (AE training code and pretrained 4096-element AEs) that would be needed to make DC-AE practically competitive with SD/SDXL VAEs.
DC-AE is an MIT / Tsinghua / NVIDIA paper about improving generative autoencoders (like the SD VAE) under the high-spatial-compression ratio regime.
mysterious software bugs I encounter, like, daily
need to eventually find repro instructions and chase all of these down to fix
| from IPython.display import HTML | |
| def get_pred_original_sample(sched, model_output, timestep, sample): | |
| return sample - sched.sigmas[(sched.timesteps == timestep).nonzero().item()] * model_output | |
| # TODO: fix awful globals | |
| prev_img_str = None | |
| def pil_to_html(pil_img, h=IM_HEIGHT, w=IM_WIDTH): | |
| global prev_img_str |
| #!/usr/bin/env python3 | |
| import gradio as gr | |
| import numpy as np | |
| import random | |
| import torch | |
| from diffusers import ( | |
| StableDiffusion3Pipeline, | |
| SD3Transformer2DModel, | |
| FlowMatchEulerDiscreteScheduler, | |
| AutoencoderTiny, |