KavinduHansaka commited on
Commit
614a574
·
verified ·
1 Parent(s): 90d6be7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +47 -5
README.md CHANGED
@@ -13,7 +13,50 @@ tags:
13
  - stable-diffusion
14
  - sdxl
15
  datasets:
16
- - KavinduHansaka/prompt-gen-8k-flux-sdxl
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  ---
18
 
19
  # Llama-3.2-1B — Image Prompt Generation (LoRA Merged)
@@ -33,12 +76,12 @@ It is designed to create **cinematic, detailed, and structured prompts** for tex
33
  - **Languages:** English (prompt tags, stylistic descriptors)
34
  - **License:** MIT
35
  - **Finetuned with:** LoRA adapters, then merged
36
- - **Training dataset:** [prompt-gen-8k-flux-sdxl](https://huggingface.co/datasets/KavinduHansaka/prompt-gen-8k-flux-sdxl)
37
 
38
  ### Model Sources
39
  - **Merged model repo:** https://huggingface.co/KavinduHansaka/Llama-3.2-1B-ImageGen
40
  - **LoRA adapter repo:** https://huggingface.co/KavinduHansaka/Llama-3.2-1B-ImageGen-LoRA
41
- - **Training dataset:** https://huggingface.co/datasets/KavinduHansaka/prompt-gen-8k-flux-sdxl
42
 
43
  ---
44
 
@@ -89,7 +132,7 @@ print(tok.decode(out[0], skip_special_tokens=True))
89
 
90
  ## Training Details
91
 
92
- - **Training data:** [prompt-gen-8k-flux-sdxl](https://huggingface.co/datasets/KavinduHansaka/prompt-gen-8k-flux-sdxl)
93
  - **Training method:** LoRA with PEFT, adapters merged into base model.
94
  - **Precision:** bfloat16/float16 during training.
95
 
@@ -111,7 +154,6 @@ title = {LLaMA 3.2 (1B)},
111
  author = {Meta AI},
112
  year = {2024},
113
  url = {https://huggingface.co/meta-llama/Llama-3.2-1B}
114
-
115
  }
116
 
117
  @misc{llama3.2-1b-promptgen,
 
13
  - stable-diffusion
14
  - sdxl
15
  datasets:
16
+ - KavinduHansaka/prompt-gen-10k-flux-sdxl
17
+
18
+ model-index:
19
+ - name: Llama-3.2-1B-ImageGen
20
+ results:
21
+ - task:
22
+ type: text-generation
23
+ name: Prompt Generation (Dev)
24
+ dataset:
25
+ type: json
26
+ name: prompt_gen_refined_dev500
27
+ metrics:
28
+ - name: eval_loss
29
+ type: loss
30
+ value: 0.7515
31
+ verified: false
32
+ - name: perplexity
33
+ type: perplexity
34
+ value: 2.12
35
+ verified: false
36
+ - name: avg_target_words
37
+ type: length
38
+ value: 106.8
39
+ verified: false
40
+
41
+ - task:
42
+ type: text-generation
43
+ name: Prompt Generation (Test)
44
+ dataset:
45
+ type: json
46
+ name: prompt_gen_refined_test500
47
+ metrics:
48
+ - name: eval_loss
49
+ type: loss
50
+ value: 0.7483
51
+ verified: false
52
+ - name: perplexity
53
+ type: perplexity
54
+ value: 2.11
55
+ verified: false
56
+ - name: avg_target_words
57
+ type: length
58
+ value: 106.7
59
+ verified: false
60
  ---
61
 
62
  # Llama-3.2-1B — Image Prompt Generation (LoRA Merged)
 
76
  - **Languages:** English (prompt tags, stylistic descriptors)
77
  - **License:** MIT
78
  - **Finetuned with:** LoRA adapters, then merged
79
+ - **Training dataset:** [prompt-gen-10k-flux-sdxl](https://huggingface.co/datasets/KavinduHansaka/prompt-gen-10k-flux-sdxl)
80
 
81
  ### Model Sources
82
  - **Merged model repo:** https://huggingface.co/KavinduHansaka/Llama-3.2-1B-ImageGen
83
  - **LoRA adapter repo:** https://huggingface.co/KavinduHansaka/Llama-3.2-1B-ImageGen-LoRA
84
+ - **Training dataset:** https://huggingface.co/datasets/KavinduHansaka/prompt-gen-10k-flux-sdxl
85
 
86
  ---
87
 
 
132
 
133
  ## Training Details
134
 
135
+ - **Training data:** [prompt-gen-8k-flux-sdxl](https://huggingface.co/datasets/KavinduHansaka/prompt-gen-10k-flux-sdxl)
136
  - **Training method:** LoRA with PEFT, adapters merged into base model.
137
  - **Precision:** bfloat16/float16 during training.
138
 
 
154
  author = {Meta AI},
155
  year = {2024},
156
  url = {https://huggingface.co/meta-llama/Llama-3.2-1B}
 
157
  }
158
 
159
  @misc{llama3.2-1b-promptgen,