python_code
stringlengths
0
456k
import logging from typing import List, Optional from .logger import DistributedLogger __all__ = ['get_dist_logger', 'DistributedLogger', 'disable_existing_loggers'] def get_dist_logger(name: str = 'colossalai') -> DistributedLogger: """Get logger instance based on name. The DistributedLogger will create single...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import inspect import logging from pathlib import Path from typing import List, Union import colossalai from colossalai.context.parallel_mode import ParallelMode class DistributedLogger: """This is a distributed event logger class essentially based on :class:`logg...
from ._trainer import Trainer __all__ = ['Trainer']
from typing import Union, List, Any import torch from torch.utils.data import DataLoader from tqdm import tqdm from colossalai.engine import Engine from colossalai.logging import DistributedLogger from colossalai.utils import MultiTimer from colossalai.utils import is_dp_rank_0, is_tp_rank_0, is_no_pp_or_last_stage f...
#!/usr/bin/env python # -*- encoding: utf-8 -*- from abc import ABC, abstractmethod from typing import Callable import torch import torch.distributed as dist from colossalai.communication import all_reduce from colossalai.context import ParallelMode from colossalai.core import global_context as gpc from colossalai.re...
#!/usr/bin/env python # -*- encoding: utf-8 -*- from abc import ABC from torch import Tensor class BaseHook(ABC): """This class allows users to add desired actions in specific time points during training or evaluation. :param priority: Priority in the printing, hooks with small priority will be printed...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import os import os.path as osp from typing import List from colossalai.context import ParallelMode from colossalai.core import global_context as gpc from colossalai.registry import HOOKS from colossalai.logging import DistributedLogger from colossalai.utils import repo...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import torch from colossalai.logging import get_dist_logger from colossalai.registry import HOOKS from colossalai.trainer.hooks import BaseHook from colossalai.utils.checkpointing import save_checkpoint from ._lr_scheduler_hook import LRSchedulerHook @HOOKS.register_mo...
from ._base_hook import BaseHook from ._checkpoint_hook import SaveCheckpointHook from ._log_hook import (LogMemoryByEpochHook, LogMetricByEpochHook, LogMetricByStepHook, LogTimingByEpochHook, TensorboardHook) from ._lr_scheduler_hook import LRSchedulerHook from ._metric_hook import AccuracyHook...
from colossalai.registry import HOOKS from torch import Tensor from ._metric_hook import LearningRateMetric, MetricHook @HOOKS.register_module class LRSchedulerHook(MetricHook): r"""Build LR scheduler for trainer. Args: lr_scheduler (:class:`colossalai.nn.lr_scheduler`): The specific LR scheduler ...
import torch def _format_number(val, prec=5): if isinstance(val, float): return f'{val:.{prec}g}' elif torch.is_tensor(val) and torch.is_floating_point(val): return f'{val.item():.{prec}g}' return val
from .cuda_native import FusedScaleMaskSoftmax, LayerNorm, MultiHeadAttention __all__ = [ "LayerNorm", "FusedScaleMaskSoftmax", "MultiHeadAttention", ]
import torch from colossalai.nn.layer.colossalai_layer import Embedding, Linear from colossalai.utils import get_current_device from .bias_dropout_add import bias_dropout_add_fused_train from .bias_gelu import bias_gelu_impl JIT_OPTIONS_SET = False def set_jit_fusion_options(): """Set PyTorch JIT layer fusion ...
from .option import set_jit_fusion_options from .bias_dropout_add import bias_dropout_add_fused_train, bias_dropout_add_fused_inference from .bias_gelu import bias_gelu_impl __all__ = [ "bias_dropout_add_fused_train", "bias_dropout_add_fused_inference", "bias_gelu_impl", "set_jit_fusion_options" ]
import torch def bias_dropout_add(x, bias, residual, prob, training): # type: (Tensor, Tensor, Tensor, float, bool) -> Tensor out = torch.nn.functional.dropout(x + bias, p=prob, training=training) out = residual + out return out @torch.jit.script def bias_dropout_add_fused_train(x: torch.Tensor, ...
import torch ###### BIAS GELU FUSION/ NO AUTOGRAD ################ # 1/sqrt(2*pi)-> 0.3989423 # 1/sqrt(2) -> 0.70710678 # sqrt(2/pi) -> 0.79788456 # this function is tanh approximation of gelu # actual gelu is: # x * 0.5 * (1.0 + torch.erf(x * 0.70710678)) @torch.jit.script def bias_gelu(bias, y): x = bias + ...
import math from dataclasses import dataclass import torch from torch import nn from torch.autograd import Function def check_config(config): if config.hidden_size % config.nhead != 0: raise Exception("hidden_size % nhead != 0") factor = 8 if config.fp16 else 4 upbound = factor * 1024 * 4 if...
from .layer_norm import MixedFusedLayerNorm as LayerNorm from .multihead_attention import MultiHeadAttention from .scaled_softmax import FusedScaleMaskSoftmax, ScaledUpperTriangMaskedSoftmax __all__ = ['LayerNorm', 'MultiHeadAttention', 'FusedScaleMaskSoftmax', 'ScaledUpperTriangMaskedSoftmax']
import enum import torch import torch.nn as nn from colossalai.kernel.op_builder.scaled_masked_softmax import ScaledMaskedSoftmaxBuilder from colossalai.kernel.op_builder.scaled_upper_triangle_masked_softmax import ScaledUpperTrainglemaskedSoftmaxBuilder try: from colossalai._C import scaled_masked_softmax, scal...
""" Fused Attention =============== This is a Triton implementation of the Flash Attention algorithm (see: Dao et al., https://arxiv.org/pdf/2205.14135v2.pdf; Rabe and Staats https://arxiv.org/pdf/2112.05682v2.pdf; Triton https://github.com/openai/triton) """ import math import os import subprocess import torch def...
"""This code is from NVIDIA apex: https://github.com/NVIDIA/apex with some changes. """ import numbers import torch from torch.cuda.amp import custom_bwd, custom_fwd from torch.nn import init from torch.nn.parameter import Parameter from colossalai.kernel.op_builder.layernorm import LayerNormBuilder try: ...
import os from torchvision.datasets import CIFAR10 def main(): dir_path = os.path.dirname(os.path.realpath(__file__)) data_root = os.path.join(dir_path, 'data') dataset = CIFAR10(root=data_root, download=True) if __name__ == '__main__': main()
import torch from torchvision.models import resnet50 from tqdm import tqdm import colossalai from colossalai.auto_parallel.tensor_shard.initialize import initialize_model from colossalai.core import global_context as gpc from colossalai.device.device_mesh import DeviceMesh from colossalai.logging import get_dist_logge...
BATCH_SIZE = 32 NUM_EPOCHS = 2
import time from copy import deepcopy from functools import partial from typing import Callable, Tuple import numpy as np import torch import torch.nn as nn import torchvision.models as tm from transformers import GPT2Config, GPT2LMHeadModel from colossalai.auto_parallel.checkpoint import CheckpointSolverRotor from c...
from setuptools import find_packages, setup setup( name='auto_parallel', version='0.0.1', description='', packages=find_packages(), install_requires=[ 'torch', 'numpy', 'tqdm', ], )
import time from argparse import ArgumentParser from functools import partial import matplotlib.pyplot as plt import torch import torch.multiprocessing as mp import torchvision.models as tm from bench_utils import GPTLMLoss, bench_rotor, data_gen_gpt2, data_gen_resnet, gpt2_medium import colossalai from colossalai.au...
import time from argparse import ArgumentParser from copy import deepcopy from functools import partial import matplotlib.pyplot as plt import numpy as np import torch import torch.multiprocessing as mp import torchvision.models as tm from bench_utils import bench, data_gen_resnet import colossalai from colossalai.au...
from colossalai.amp import AMP_TYPE # hyper-parameters TRAIN_ITERS = 10 DECAY_ITERS = 4 WARMUP_FRACTION = 0.01 GLOBAL_BATCH_SIZE = 32 # dp world size * sentences per GPU EVAL_ITERS = 10 EVAL_INTERVAL = 10 LR = 0.0001 MIN_LR = 1e-05 WEIGHT_DECAY = 0.01 SEQ_LENGTH = 128 # BERT config DEPTH = 4 NUM_ATTENTION_HEADS = ...
import argparse import torch from data.bert_helper import SequenceParallelDataIterator, get_batch_for_sequence_parallel from data.dummy_dataloader import DummyDataloader from loss_func.bert_loss import BertLoss from lr_scheduler import AnnealingLR from model.bert import BertForPretrain, build_pipeline_bert import col...
from .annealing_lr import AnnealingLR
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
from colossalai.context.parallel_mode import ParallelMode import torch from torch.cuda.amp import custom_bwd, custom_fwd class _VocabCrossEntropy(torch.autograd.Function): @staticmethod @custom_fwd def forward(ctx, vocab_parallel_logits, target): # Maximum value along vocab dimension across all G...
import torch import torch.nn as nn from colossalai.core import global_context as gpc from colossalai.context import ParallelMode from colossalai.logging import get_dist_logger import torch.nn.functional as F import torch.distributed as dist from .cross_entropy import vocab_cross_entropy class BertLoss(nn.Module): ...
import torch def ensure_divisibility(numerator, denominator): """Ensure that numerator is divisible by the denominator.""" assert numerator % denominator == 0, '{} is not divisible by {}'.format( numerator, denominator) def divide(numerator, denominator): """Ensure that numerator is divisible b...
from colossalai.context.parallel_mode import ParallelMode import torch import torch.nn as nn import inspect from .layers import Embedding, BertLayer, BertDualHead, PreProcessor, VocabEmbedding from .layers.init_method import init_normal, output_init_normal from colossalai.core import global_context as gpc from colossal...
from colossalai.context.parallel_mode import ParallelMode import torch import torch.nn as nn from colossalai.core import global_context as gpc class PreProcessor(nn.Module): def __init__(self, sub_seq_length): super().__init__() self.sub_seq_length = sub_seq_length def bert_position_ids(self...
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.init as init class VocabEmbedding(torch.nn.Module): def __init__(self, num_embeddings, embedding_dim): super(VocabEmbedding, self).__init__() # Keep the input dimensions. self.num_embeddings = num_embedding...
import torch import torch.nn as nn from torch.nn import Parameter import torch.nn.functional as F import torch.nn.init as init class Linear(nn.Module): """Linear layer with column parallelism. The linear layer is defined as Y = XA + b. A is parallelized along its second dimension as A = [A_1, ..., A_p]. ...
import torch import torch.nn as nn from colossalai.nn.layer.parallel_sequence import TransformerSelfAttentionRing from colossalai.kernel.jit import bias_dropout_add_fused_train, bias_dropout_add_fused_inference from colossalai.kernel.cuda_native import LayerNorm from .mlp import TransformerMLP from .dropout import get_...
from .embedding import VocabEmbedding, Embedding from .bert_layer import BertLayer from .head import BertDualHead from .preprocess import PreProcessor
import torch import torch.nn as nn import torch.nn.functional as F from .linear import Linear from colossalai.kernel.jit import bias_gelu_impl class TransformerMLP(nn.Module): """MLP. MLP will take the input with h hidden state, project it to 4*h hidden dimension, perform nonlinear transformation, and pr...
import torch import math def init_normal(tensor, sigma): """Init method based on N(0, sigma).""" torch.nn.init.normal_(tensor, mean=0.0, std=sigma) def output_init_normal(tensor, sigma, num_layers): """Init method based on N(0, sigma/sqrt(2*num_layers).""" std = sigma / math.sqrt(2.0 * num_layers) ...
import colossalai import torch import torch.nn as nn import torch.nn.functional as F from .pooler import Pooler from .linear import Linear from .embedding import VocabEmbedding from colossalai.core import global_context as gpc from colossalai.context import ParallelMode from colossalai.kernel import LayerNorm from loss...
import torch def bias_dropout_add(x, bias, residual, prob, training): # type: (Tensor, Tensor, Tensor, float, bool) -> Tensor out = torch.nn.functional.dropout(x + bias, p=prob, training=training) out = residual + out return out def get_bias_dropout_add(training): def _bias_dropout_add(x, bias, r...
import torch import torch.nn as nn from .linear import Linear class Pooler(nn.Module): """Pooler layer. Pool hidden states of a specific token (for example start of the sequence) and add a linear transformation followed by a tanh. Arguments: hidden_size: hidden size init_method: weig...
from colossalai.core import global_context as gpc from colossalai.context import ParallelMode import torch _MAX_DATA_DIM = 5 def _build_key_size_numel_dictionaries(keys, data): """Build the size on rank 0 and broadcast.""" max_dim = _MAX_DATA_DIM sizes = [0 for _ in range(max_dim) for _ in keys] # P...
import torch class DummyDataloader(): def __init__(self, batch_size, vocab_size, seq_length): self.batch_size = batch_size self.vocab_size = vocab_size self.seq_length = seq_length self.step = 0 def generate(self): tokens = torch.randint(low=0, high=self.vocab_size, s...
from colossalai.context.parallel_context import ParallelContext from colossalai.core import global_context as gpc from colossalai.logging import get_dist_logger from colossalai.context import ParallelMode from .datasets.data_samplers import build_pretraining_data_loader from .datasets.builder import build_train_valid_t...
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by ...
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
from . import indexed_dataset
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors, and NVIDIA. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless ...
from .blendable_dataset import BlendableDataset from .dataset_utils import get_datasets_weights_and_num_samples, get_indexed_dataset_, get_train_valid_test_split_ from .bert_dataset import BertDataset from colossalai.logging import get_dist_logger DSET_TYPE_BERT = 'standard_bert' DSET_TYPE_ICT = 'ict' DSET_TYPE_T5 = '...
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # copied from fairseq/fairseq/data/indexed_dataset.py # Removed IndexedRawTextDataset since it relied on Fairseq dictionary # other slight mo...
import itertools import random import numpy as np from torch.utils.data import Dataset from megatron import get_tokenizer from megatron import get_args from megatron.data.dataset_utils import get_indexed_dataset_ from megatron.data.realm_dataset_utils import get_block_samples_mapping def make_attention_mask(source_b...
# This file isn't really a formal automated test, it's just a place to # put some code used during development and manual testing of # indexed_dataset. from megatron.data import indexed_dataset from megatron.tokenizer import build_tokenizer import argparse import os import sys import torch script_dir = os.path.dirna...
from colossalai.amp import AMP_TYPE # hyperparameters # BATCH_SIZE is as per GPU # global batch size = BATCH_SIZE x data parallel size BATCH_SIZE = 4 LEARNING_RATE = 3e-3 WEIGHT_DECAY = 0.3 NUM_EPOCHS = 2 WARMUP_EPOCHS = 1 # model config IMG_SIZE = 224 PATCH_SIZE = 16 HIDDEN_SIZE = 128 DEPTH = 4 NUM_HEADS = 4 MLP_RAT...
import os import torch from titans.model.vit.vit import _create_vit_model from tqdm import tqdm import colossalai from colossalai.context import ParallelMode from colossalai.core import global_context as gpc from colossalai.logging import get_dist_logger from colossalai.nn import CrossEntropyLoss from colossalai.nn.l...
import argparse import logging import random from typing import Optional import uvicorn from energonai import QueueFullError, launch_engine from energonai.model import opt_6B, opt_30B, opt_125M, opt_175B from fastapi import FastAPI, HTTPException, Request from pydantic import BaseModel, Field from transformers import ...
import torch from typing import List, Deque, Tuple, Hashable, Any from energonai import BatchManager, SubmitEntry, TaskEntry class BatchManagerForGeneration(BatchManager): def __init__(self, max_batch_size: int = 1, pad_token_id: int = 0) -> None: super().__init__() self.max_batch_size = max_batch...
from collections import OrderedDict from threading import Lock from contextlib import contextmanager from typing import List, Any, Hashable, Dict class MissCacheError(Exception): pass class ListCache: def __init__(self, cache_size: int, list_size: int, fixed_keys: List[Hashable] = []) -> None: """Ca...
import logging import argparse import random from torch import Tensor from pydantic import BaseModel, Field from typing import Optional from energonai.model import opt_125M, opt_30B, opt_175B, opt_6B from transformers import GPT2Tokenizer from energonai import launch_engine, QueueFullError from sanic import Sanic from ...
from locust import HttpUser, task from json import JSONDecodeError class GenerationUser(HttpUser): @task def generate(self): prompt = 'Question: What is the longest river on the earth? Answer:' for i in range(4, 9): data = {'max_tokens': 2**i, 'prompt': prompt} with sel...
import os import torch from multiprocessing import Pool # download pytorch model ckpt in https://huggingface.co/facebook/opt-66b/tree/main # you can use whether wget or git lfs path = "/path/to/your/ckpt" new_path = "/path/to/the/processed/ckpt/" assert os.path.isdir(path) files = [] for filename in os.listdir(path)...
import argparse import json import os import re from collections import defaultdict import numpy as np import torch def load_json(path: str): with open(path) as f: return json.load(f) def parse_shape_info(flat_dir: str): data = load_json(os.path.join(flat_dir, 'shape.json')) flat_info = default...
#!/usr/bin/env python # coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
from colossalai.zero.shard_utils import TensorShardStrategy zero = dict(model_config=dict(shard_strategy=TensorShardStrategy(), tensor_placement_policy="auto", reuse_fp16_shard=True), optimizer_config=dict(gpu_margin_mem_ratio=0.8, initial_scale=1...
import torch.distributed as dist from colossalai.context import ParallelMode from colossalai.core import global_context as gpc class barrier_context(): """ This context manager is used to allow one process to execute while blocking all other processes in the same process group. This is often useful when ...
from colossalai.amp import AMP_TYPE # hyperparameters # BATCH_SIZE is as per GPU # global batch size = BATCH_SIZE x data parallel size BATCH_SIZE = 512 LEARNING_RATE = 3e-3 WEIGHT_DECAY = 0.3 NUM_EPOCHS = 2 WARMUP_EPOCHS = 1 # model config NUM_CLASSES = 10 fp16 = dict(mode=AMP_TYPE.NAIVE) clip_grad_norm = 1.0
import torch import torch.nn as nn from torchvision.models import resnet18 from tqdm import tqdm import colossalai from colossalai.core import global_context as gpc from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import Lamb, La...
from setuptools import setup, find_packages setup( name='latent-diffusion', version='0.0.1', description='', packages=find_packages(), install_requires=[ 'torch', 'numpy', 'tqdm', ], )
import argparse import csv import datetime import glob import importlib import os import sys import time import numpy as np import torch import torchvision try: import lightning.pytorch as pl except: import pytorch_lightning as pl from functools import partial from omegaconf import OmegaConf from packaging ...
import argparse, os, sys, glob from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm import numpy as np import torch from main import instantiate_from_config from ldm.models.diffusion.ddim import DDIMSampler def make_batch(image, mask, device): image = np.array(Image.open(image).convert("RGB...
import argparse, os import cv2 import torch import numpy as np from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange from torchvision.utils import make_grid try: from lightning.pytorch import seed_everything except: from pyt...
import argparse, os, sys, glob import clip import torch import torch.nn as nn import numpy as np from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid import scann import time from mul...
"""make variations of input image""" import argparse, os import PIL import torch import numpy as np from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast fr...
import bitsandbytes as bnb import torch.nn as nn import torch class Linear8bit(nn.Linear): def __init__( self, input_features, output_features, bias=True, has_fp16_weights=False, memory_efficient_backward=False, threshold=6.0, weight_data=None, ...
import argparse, os, sys, glob, datetime, yaml import torch import time import numpy as np from tqdm import trange from omegaconf import OmegaConf from PIL import Image from ldm.models.diffusion.ddim import DDIMSampler from ldm.util import instantiate_from_config rescale = lambda x: (x + 1.) / 2. def custom_to_pil(...
import os, sys import numpy as np import scann import argparse import glob from multiprocessing import cpu_count from tqdm import tqdm from ldm.util import parallel_data_prefetch def search_bruteforce(searcher): return searcher.score_brute_force().build() def search_partioned_ah(searcher, dims_per_block, aiq_t...
import cv2 import fire from imwatermark import WatermarkDecoder def testit(img_path): bgr = cv2.imread(img_path) decoder = WatermarkDecoder('bytes', 136) watermark = decoder.decode(bgr, 'dwtDct') try: dec = watermark.decode('utf-8') except: dec = "null" print(dec) if __name__...
import os import sys from copy import deepcopy import yaml from datetime import datetime from diffusers import StableDiffusionPipeline import torch from ldm.util import instantiate_from_config from main import get_parser if __name__ == "__main__": with torch.no_grad(): yaml_path = "../../train_colossalai...
import numpy as np class LambdaWarmUpCosineScheduler: """ note: use with a base_lr of 1.0 """ def __init__(self, warm_up_steps, lr_min, lr_max, lr_start, max_decay_steps, verbosity_interval=0): self.lr_warm_up_steps = warm_up_steps self.lr_start = lr_start self.lr_min = lr_min ...
import importlib import torch from torch import optim import numpy as np from inspect import isfunction from PIL import Image, ImageDraw, ImageFont def log_txt_as_img(wh, xc, size=10): # wh a tuple of (width, height) # xc a list of captions to plot b = len(xc) txts = list() for bi in range(b): ...
import torch try: import lightning.pytorch as pl except: import pytorch_lightning as pl import torch.nn.functional as F from contextlib import contextmanager from ldm.modules.diffusionmodules.model import Encoder, Decoder from ldm.modules.distributions.distributions import DiagonalGaussianDistribution from l...
"""SAMPLING ONLY.""" import torch import numpy as np from tqdm import tqdm from ldm.modules.diffusionmodules.util import make_ddim_sampling_parameters, make_ddim_timesteps, noise_like, extract_into_tensor class DDIMSampler(object): def __init__(self, model, schedule="linear", **kwargs): super().__init__...
import torch import numpy as np def append_dims(x, target_dims): """Appends dimensions to the end of a tensor until it has target_dims dimensions. From https://github.com/crowsonkb/k-diffusion/blob/master/k_diffusion/utils.py""" dims_to_append = target_dims - x.ndim if dims_to_append < 0: rais...
import os import torch import lightning.pytorch as pl from omegaconf import OmegaConf from torch.nn import functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from copy import deepcopy from einops import rearrange from glob import glob from natsort import natsorted from ldm.modu...
"""SAMPLING ONLY.""" import torch import numpy as np from tqdm import tqdm from functools import partial from ldm.modules.diffusionmodules.util import make_ddim_sampling_parameters, make_ddim_timesteps, noise_like from ldm.models.diffusion.sampling_util import norm_thresholding class PLMSSampler(object): def __...
""" wild mixture of https://github.com/lucidrains/denoising-diffusion-pytorch/blob/7706bdfc6f527f58d33f84b7b522e61e6e3164b3/denoising_diffusion_pytorch/denoising_diffusion_pytorch.py https://github.com/openai/improved-diffusion/blob/e94489283bb876ac1477d5dd7709bbbd2d9902ce/improved_diffusion/gaussian_diffusion.py https...
import torch import torch.nn.functional as F import math from tqdm import tqdm class NoiseScheduleVP: def __init__( self, schedule='discrete', betas=None, alphas_cumprod=None, continuous_beta_0=0.1, continuous_beta_1=20., ): """Cr...
from .sampler import DPMSolverSampler
"""SAMPLING ONLY.""" import torch from .dpm_solver import NoiseScheduleVP, model_wrapper, DPM_Solver MODEL_TYPES = { "eps": "noise", "v": "v" } class DPMSolverSampler(object): def __init__(self, model, **kwargs): super().__init__() self.model = model to_torch = lambda x: x.clone...