RectifiedFlowLM / config.json
pbansal's picture
Upload folder using huggingface_hub
3cccc0a verified
raw
history blame contribute delete
349 Bytes
{
"architectures": [
"DIT"
],
"cond_dim": 128,
"dropout": 0.1,
"embedding_size": 768,
"hidden_size": 768,
"length": 1024,
"model_type": "dit",
"n_blocks": 12,
"n_heads": 12,
"name": "small",
"noise_num_pieces": 16,
"torch_dtype": "float32",
"transformers_version": "4.51.3",
"type": "ddit",
"vocab_size": 50258
}