File size: 374 Bytes
fac61f6
 
 
 
 
 
 
 
 
 
 
 
 
 
de9a113
fac61f6
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
import torch

# NOTE model path name changed
model_path = './last_epoch_ckpt/'
lr = 1e-5
device = 'cuda'
dtype = torch.bfloat16
data_path = '../data/lke_2017'
save_path = './'
epochs = 4
batch_size = 16
number_k_clip_embed = 16 # divide by this to determine bundling together of sequences -> CLIP
num_workers = 32
seed = 107
k = 8
# TODO config option to swap to diffusion?