Skip to content

Instantly share code, notes, and snippets.

View a-r-r-o-w's full-sized avatar
:octocat:
wandering on a rock

Aryan V S a-r-r-o-w

:octocat:
wandering on a rock
View GitHub Profile
@a-r-r-o-w
a-r-r-o-w / cogvideox-quanto-inference.py
Last active November 26, 2024 02:20
Demonstrates how to use CogVideoX 2B/5B with Diffusers and Optimum-Quanto
import gc
import torch
from diffusers import CogVideoXPipeline
from diffusers.utils import export_to_video
from optimum.quanto import freeze, quantize, qfloat8, qfloat8_e4m3fn, qfloat8_e5m2, qint8, qint4, qint2
def reset_memory(device):
gc.collect()
torch.cuda.empty_cache()
@a-r-r-o-w
a-r-r-o-w / cogvideo-torchao-compile.py
Created August 20, 2024 11:43
Demonstrates CogVideoX quantized WO-inference with torchao
import argparse
import gc
import os
import time
os.environ["TORCH_LOGS"] = "dynamo"
import torch
from diffusers import AutoencoderKLCogVideoX, CogVideoXPipeline, CogVideoXTransformer3DModel, CogVideoXDDIMScheduler
from diffusers.utils import export_to_video
import gc
import torch
from diffusers import CogVideoXPipeline, CogVideoXDDIMScheduler
from diffusers.utils import export_to_video
def reset_memory():
gc.collect()
torch.cuda.empty_cache()
import argparse
import time
import torch
torch.set_float32_matmul_precision("high")
from diffusers import CogVideoXPipeline, CogVideoXDDIMScheduler
from diffusers.utils import export_to_gif
import argparse
import time
import torch
torch.set_float32_matmul_precision("high")
from diffusers import LattePipeline
from diffusers.utils import export_to_gif