51 lines
1.4 KiB
Markdown
51 lines
1.4 KiB
Markdown
|
---
|
||
|
{{ card_data }}
|
||
|
---
|
||
|
|
||
|
<!-- This model card has been generated automatically according to the information the training script had access to. You
|
||
|
should probably proofread and complete it, then remove this comment. -->
|
||
|
|
||
|
# {{ model_name | default("Diffusion Model") }}
|
||
|
|
||
|
## Model description
|
||
|
|
||
|
This diffusion model is trained with the [🤗 Diffusers](https://github.com/huggingface/diffusers) library
|
||
|
on the `{{ dataset_name }}` dataset.
|
||
|
|
||
|
## Intended uses & limitations
|
||
|
|
||
|
#### How to use
|
||
|
|
||
|
```python
|
||
|
# TODO: add an example code snippet for running this diffusion pipeline
|
||
|
```
|
||
|
|
||
|
#### Limitations and bias
|
||
|
|
||
|
[TODO: provide examples of latent issues and potential remediations]
|
||
|
|
||
|
## Training data
|
||
|
|
||
|
[TODO: describe the data used to train the model]
|
||
|
|
||
|
### Training hyperparameters
|
||
|
|
||
|
The following hyperparameters were used during training:
|
||
|
- learning_rate: {{ learning_rate }}
|
||
|
- train_batch_size: {{ train_batch_size }}
|
||
|
- eval_batch_size: {{ eval_batch_size }}
|
||
|
- gradient_accumulation_steps: {{ gradient_accumulation_steps }}
|
||
|
- optimizer: AdamW with betas=({{ adam_beta1 }}, {{ adam_beta2 }}), weight_decay={{ adam_weight_decay }} and epsilon={{ adam_epsilon }}
|
||
|
- lr_scheduler: {{ lr_scheduler }}
|
||
|
- lr_warmup_steps: {{ lr_warmup_steps }}
|
||
|
- ema_inv_gamma: {{ ema_inv_gamma }}
|
||
|
- ema_inv_gamma: {{ ema_power }}
|
||
|
- ema_inv_gamma: {{ ema_max_decay }}
|
||
|
- mixed_precision: {{ mixed_precision }}
|
||
|
|
||
|
### Training results
|
||
|
|
||
|
📈 [TensorBoard logs](https://huggingface.co/{{ repo_name }}/tensorboard?#scalars)
|
||
|
|
||
|
|