tkarras commited on
Commit
d573e4a
·
verified ·
1 Parent(s): 778e927

Add files using upload-large-folder tool

Browse files
Files changed (50) hide show
  1. LICENSE.txt +86 -0
  2. README.md +92 -4
  3. fdfo-vlm-alignment-reward-no-cfg/epoch-0000005/adapter_config.json +50 -0
  4. fdfo-vlm-alignment-reward-no-cfg/epoch-0000010/adapter_config.json +50 -0
  5. fdfo-vlm-alignment-reward-no-cfg/epoch-0000015/adapter_config.json +50 -0
  6. fdfo-vlm-alignment-reward-no-cfg/epoch-0000020/adapter_config.json +50 -0
  7. fdfo-vlm-alignment-reward-no-cfg/epoch-0000025/adapter_config.json +50 -0
  8. fdfo-vlm-alignment-reward-no-cfg/epoch-0000030/adapter_config.json +50 -0
  9. fdfo-vlm-alignment-reward-no-cfg/epoch-0000035/adapter_config.json +50 -0
  10. fdfo-vlm-alignment-reward-no-cfg/epoch-0000040/adapter_config.json +50 -0
  11. fdfo-vlm-alignment-reward-no-cfg/epoch-0000045/adapter_config.json +50 -0
  12. fdfo-vlm-alignment-reward-no-cfg/epoch-0000050/adapter_config.json +50 -0
  13. fdfo-vlm-alignment-reward-no-cfg/epoch-0000055/adapter_config.json +50 -0
  14. fdfo-vlm-alignment-reward-no-cfg/epoch-0000060/adapter_config.json +50 -0
  15. fdfo-vlm-alignment-reward-no-cfg/epoch-0000065/adapter_config.json +50 -0
  16. fdfo-vlm-alignment-reward-no-cfg/epoch-0000070/adapter_config.json +50 -0
  17. fdfo-vlm-alignment-reward-no-cfg/epoch-0000075/adapter_config.json +50 -0
  18. fdfo-vlm-alignment-reward-no-cfg/epoch-0000080/adapter_config.json +50 -0
  19. fdfo-vlm-alignment-reward-no-cfg/epoch-0000085/adapter_config.json +50 -0
  20. fdfo-vlm-alignment-reward-no-cfg/epoch-0000090/adapter_config.json +50 -0
  21. fdfo-vlm-alignment-reward-no-cfg/epoch-0000095/adapter_config.json +50 -0
  22. fdfo-vlm-alignment-reward-no-cfg/epoch-0000100/adapter_config.json +50 -0
  23. fdfo-vlm-alignment-reward-no-cfg/epoch-0000105/adapter_config.json +50 -0
  24. fdfo-vlm-alignment-reward-no-cfg/epoch-0000110/adapter_config.json +50 -0
  25. fdfo-vlm-alignment-reward-no-cfg/epoch-0000115/adapter_config.json +50 -0
  26. fdfo-vlm-alignment-reward-no-cfg/epoch-0000120/adapter_config.json +50 -0
  27. fdfo-vlm-alignment-reward-no-cfg/epoch-0000125/adapter_config.json +50 -0
  28. fdfo-vlm-alignment-reward-no-cfg/epoch-0000130/adapter_config.json +50 -0
  29. fdfo-vlm-alignment-reward-no-cfg/epoch-0000135/adapter_config.json +50 -0
  30. fdfo-vlm-alignment-reward-no-cfg/epoch-0000140/adapter_config.json +50 -0
  31. fdfo-vlm-alignment-reward-no-cfg/epoch-0000145/adapter_config.json +50 -0
  32. fdfo-vlm-alignment-reward-no-cfg/epoch-0000150/adapter_config.json +50 -0
  33. fdfo-vlm-alignment-reward-no-cfg/epoch-0000155/adapter_config.json +50 -0
  34. fdfo-vlm-alignment-reward-no-cfg/epoch-0000160/adapter_config.json +50 -0
  35. fdfo-vlm-alignment-reward-no-cfg/epoch-0000165/adapter_config.json +50 -0
  36. fdfo-vlm-alignment-reward-no-cfg/epoch-0000170/adapter_config.json +50 -0
  37. fdfo-vlm-alignment-reward-no-cfg/epoch-0000175/adapter_config.json +50 -0
  38. fdfo-vlm-alignment-reward-no-cfg/epoch-0000180/adapter_config.json +50 -0
  39. fdfo-vlm-alignment-reward-no-cfg/epoch-0000185/adapter_config.json +50 -0
  40. fdfo-vlm-alignment-reward-no-cfg/epoch-0000190/adapter_config.json +50 -0
  41. fdfo-vlm-alignment-reward-no-cfg/epoch-0000195/adapter_config.json +50 -0
  42. fdfo-vlm-alignment-reward-no-cfg/epoch-0000450/adapter_config.json +50 -0
  43. fdfo-vlm-alignment-reward-no-cfg/epoch-0000455/adapter_config.json +50 -0
  44. fdfo-vlm-alignment-reward-no-cfg/epoch-0000460/adapter_config.json +50 -0
  45. fdfo-vlm-alignment-reward-no-cfg/epoch-0000465/adapter_config.json +50 -0
  46. fdfo-vlm-alignment-reward-no-cfg/epoch-0000470/adapter_config.json +50 -0
  47. fdfo-vlm-alignment-reward-no-cfg/epoch-0000475/adapter_config.json +50 -0
  48. fdfo-vlm-alignment-reward-no-cfg/epoch-0000480/adapter_config.json +50 -0
  49. fdfo-vlm-alignment-reward-no-cfg/epoch-0000485/adapter_config.json +50 -0
  50. fdfo-vlm-alignment-reward-no-cfg/epoch-0000490/adapter_config.json +50 -0
LICENSE.txt ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ NVIDIA Source Code License v1 (Non-Commercial)
2
+ ==============================================
3
+
4
+ NVIDIA License
5
+
6
+ 1. Definitions
7
+
8
+ "Licensor" means any person or entity that distributes its Work.
9
+
10
+ "Work" means (a) the original work of authorship made available under this
11
+ license, which may include software, documentation, or other files, and
12
+ (b) any additions to or derivative works thereof that are made available
13
+ under this license.
14
+
15
+ The terms "reproduce," "reproduction," "derivative works," and "distribution"
16
+ have the meaning as provided under U.S. copyright law; provided, however,
17
+ that for the purposes of this license, derivative works shall not include
18
+ works that remain separable from, or merely link (or bind by name) to the
19
+ interfaces of, the Work.
20
+
21
+ Works are "made available" under this license by including in or with the
22
+ Work either (a) a copyright notice referencing the applicability of this
23
+ license to the Work, or (b) a copy of this license.
24
+
25
+ 2. License Grant
26
+
27
+ 2.1 Copyright Grant. Subject to the terms and conditions of this license,
28
+ each Licensor grants to you a perpetual, worldwide, non-exclusive,
29
+ royalty-free, copyright license to use, reproduce, prepare derivative works
30
+ of, publicly display, publicly perform, sublicense and distribute its Work
31
+ and any resulting derivative works in any form.
32
+
33
+ 3. Limitations
34
+
35
+ 3.1 Redistribution. You may reproduce or distribute the Work only if (a) you
36
+ do so under this license, (b) you include a complete copy of this license
37
+ with your distribution, and (c) you retain without modification any
38
+ copyright, patent, trademark, or attribution notices that are present in the
39
+ Work.
40
+
41
+ 3.2 Derivative Works. You may specify that additional or different terms
42
+ apply to the use, reproduction, and distribution of your derivative works of
43
+ the Work ("Your Terms") only if (a) Your Terms provide that the use
44
+ limitation in Section 3.3 applies to your derivative works, and (b) you
45
+ identify the specific derivative works that are subject to Your Terms.
46
+ Notwithstanding Your Terms, this license (including the redistribution
47
+ requirements in Section 3.1) will continue to apply to the Work itself.
48
+
49
+ 3.3 Use Limitation. The Work and any derivative works thereof only may be
50
+ used or intended for use non-commercially. Notwithstanding the foregoing,
51
+ NVIDIA Corporation and its affiliates may use the Work and any derivative
52
+ works commercially. As used herein, "non-commercially" means for research or
53
+ evaluation purposes only.
54
+
55
+ 3.4 Patent Claims. If you bring or threaten to bring a patent claim against
56
+ any Licensor (including any claim, cross-claim or counterclaim in a lawsuit)
57
+ to enforce any patents that you allege are infringed by any Work, then your
58
+ rights under this license from such Licensor (including the grant in Section
59
+ 2.1) will terminate immediately.
60
+
61
+ 3.5 Trademarks. This license does not grant any rights to use any Licensor's
62
+ or its affiliates' names, logos, or trademarks, except as necessary to
63
+ reproduce the notices described in this license.
64
+
65
+ 3.6 Termination. If you violate any term of this license, then your rights
66
+ under this license (including the grant in Section 2.1) will terminate
67
+ immediately.
68
+
69
+ 4. Disclaimer of Warranty.
70
+
71
+ THE WORK IS PROVIDED "AS IS" WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
72
+ EITHER EXPRESS OR IMPLIED, INCLUDING WARRANTIES OR CONDITIONS OF
73
+ MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, TITLE OR
74
+ NON-INFRINGEMENT. YOU BEAR THE RISK OF UNDERTAKING ANY ACTIVITIES UNDER THIS
75
+ LICENSE.
76
+
77
+ 5. Limitation of Liability.
78
+
79
+ EXCEPT AS PROHIBITED BY APPLICABLE LAW, IN NO EVENT AND UNDER NO LEGAL
80
+ THEORY, WHETHER IN TORT (INCLUDING NEGLIGENCE), CONTRACT, OR OTHERWISE SHALL
81
+ ANY LICENSOR BE LIABLE TO YOU FOR DAMAGES, INCLUDING ANY DIRECT, INDIRECT,
82
+ SPECIAL, INCIDENTAL, OR CONSEQUENTIAL DAMAGES ARISING OUT OF OR RELATED TO
83
+ THIS LICENSE, THE USE OR INABILITY TO USE THE WORK (INCLUDING BUT NOT LIMITED
84
+ TO LOSS OF GOODWILL, BUSINESS INTERRUPTION, LOST PROFITS OR DATA, COMPUTER
85
+ FAILURE OR MALFUNCTION, OR ANY OTHER DAMAGES OR LOSSES), EVEN IF THE LICENSOR
86
+ HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES.
README.md CHANGED
@@ -1,4 +1,92 @@
1
- ---
2
- license: unknown
3
- ---
4
- # TODO
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # FDFO: Finite Difference Flow Optimization
2
+
3
+ This repository contains the official pretrained checkpoints for FDFO, a method for fine-tuning flow-based diffusion models using finite difference gradient estimation. We fine-tune [Stable Diffusion 3.5 Medium](https://huggingface.co/stabilityai/stable-diffusion-3.5-medium) using reward signals from VLM-based scoring and/or PickScore. Please see the [GitHub repository](https://github.com/NVlabs/finite-difference-flow-optimization) for further details.
4
+
5
+ Powered by Stability AI.
6
+
7
+ # Model overview
8
+
9
+ ## Description
10
+ Finite Difference Flow Optimization is an algorithm for reinforcement learning (RL) post-training of text-to-image diffusion models to improve output quality. We provide a set of Low-Rank Adaptation (LoRA) checkpoints that modify the behavior of the off-the-shelf Stable Diffusion 3.5 Medium text-to-image model.
11
+
12
+ This model is for research and development only.
13
+
14
+ ## License
15
+ The FDFO source code and pretrained checkpoints are licensed under the [NVIDIA Source Code License v1 (Non-Commercial)](LICENSE.txt). Copyright © 2026 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
16
+
17
+ The Stable Diffusion 3.5 Medium Model is licensed under the [Stability AI Community License](https://stability.ai/community-license-agreement). Copyright © Stability AI Ltd. All rights reserved.
18
+
19
+ ## Deployment geography
20
+ Global
21
+
22
+ ## Use case
23
+ The checkpoints are intended for academic researchers who want to reproduce the results from the following NVIDIA research paper.
24
+
25
+ **Finite Difference Flow Optimization for RL Post-Training of Text-to-Image Models**<br/>
26
+ David McAllister, Miika Aittala, Tero Karras, Janne Hellsten, Angjoo Kanazawa, Timo Aila, Samuli Laine<br/>
27
+ https://arxiv.org/abs/TODO.TODO
28
+
29
+ ## Release date
30
+ TODO
31
+
32
+ ## References
33
+ **Research paper:** https://arxiv.org/abs/TODO.TODO<br/>
34
+ **Source code:** https://github.com/NVlabs/finite-difference-flow-optimization<br/>
35
+ **Checkpoints:** https://huggingface.co/nvidia/finite-difference-flow-optimization<br/>
36
+
37
+ ## Model architecture
38
+ **Architecture type:** Transformer LoRA<br/>
39
+ **Network architecture:** Low-rank adapter for Stable Diffusion 3.5 Medium<br/>
40
+ **Number of model parameters:** 1.9*10^7<br/>
41
+
42
+ The low-rank adapter was initialized to zero and trained using Finite Difference Flow Optimization for 1000 RL epochs, where one RL epoch corresponds to 864 reward evaluations. See the associated [research paper](https://arxiv.org/abs/TODO.TODO) for further details.
43
+
44
+ ## Input
45
+ **Input type:** Text<br/>
46
+ **Input format:** String<br/>
47
+ **Input parameters:** One-dimensional (1D)<br/>
48
+ **Other properties related to input:** Typically around 50-200 letters, up to 2K tokens<br/>
49
+
50
+ ## Output
51
+ **Output type:** Image<br/>
52
+ **Output format:** Red, green, blue (RGB)<br/>
53
+ **Output parameters:** Two-dimensional (2D)<br/>
54
+ **Other properties related to output:** 512x512, 24 bits per pixel<br/>
55
+
56
+ ## Software integration
57
+ **Runtime engine:** PyTorch 2.7.0<br/>
58
+ **Supported hardware microarchitecture compatibility:** NVIDIA Ampere, NVIDIA Hopper<br/>
59
+ **Preferred/supported operating system:** Linux<br/>
60
+
61
+ Our AI models are designed and/or optimized to run on NVIDIA GPU-accelerated systems. By leveraging NVIDIA’s hardware (e.g. GPU cores) and software frameworks (e.g., CUDA libraries), the model achieves faster training and inference times compared to CPU-only solutions.
62
+
63
+ The integration of foundation and fine-tuned models into AI systems requires additional testing using use-case-specific data to ensure safe and effective deployment. Following the V-model methodology, iterative testing and validation at both unit and system levels are essential to mitigate risks, meet technical and functional requirements, and ensure compliance with safety and ethical standards before deployment.
64
+
65
+ ## Model version
66
+ 1.0
67
+
68
+ ## Training and evaluation datasets
69
+
70
+ **Training dataset:** Pick-a-Pic<br/>
71
+ **Link:** https://stability.ai/research/pick-a-pic<br/>
72
+ **Data modality:** Text prompts (25432 prompts)<br/>
73
+ **Data collection method:** Human<br/>
74
+ Web-based human preference collection run by Stability AI via the Pick-a-Pic interface with explicit user consent.<br/>
75
+ **Text training data size:** Less than a Billion Tokens<br/>
76
+ **Labeling method:** Not applicable<br/>
77
+
78
+ **Evaluation dataset:** HPDv2<br/>
79
+ **Link:** https://huggingface.co/datasets/ymhao/HPDv2<br/>
80
+ **Data collection method:** Hybrid (Human, Synthetic)<br/>
81
+ The data consists of text prompts from MS COCO and DiffusionDB, cleaned by ChatGPT. The dataset collection method is described in detail in the [Human Preference Score v2](https://arxiv.org/abs/2306.09341) paper.<br/>
82
+ **Properties:** 3200 text prompts (evaluation set)<br/>
83
+ **Labeling method:** Not applicable<br/>
84
+
85
+ ## Inference
86
+ **Acceleration engine:** PyTorch 2.7.0<br/>
87
+ **Test hardware:** NVIDIA Hopper<br/>
88
+
89
+ ## Ethical considerations
90
+ NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
91
+
92
+ Please report model quality, risk, security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).
fdfo-vlm-alignment-reward-no-cfg/epoch-0000005/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000010/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000015/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000020/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000025/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000030/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000035/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000040/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000045/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000050/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000055/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000060/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000065/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000070/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000075/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000080/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000085/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000090/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000095/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000100/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000105/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000110/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000115/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000120/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000125/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000130/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000135/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000140/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000145/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000150/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000155/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000160/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000165/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000170/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000175/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000180/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000185/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000190/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000195/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000450/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000455/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000460/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000465/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000470/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000475/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000480/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000485/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
fdfo-vlm-alignment-reward-no-cfg/epoch-0000490/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "SD3Transformer2DModel",
7
+ "parent_library": "diffusers.models.transformers.transformer_sd3"
8
+ },
9
+ "base_model_name_or_path": null,
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": "gaussian",
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 64,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.0,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 32,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "attn.to_add_out",
36
+ "attn.add_k_proj",
37
+ "attn.to_out.0",
38
+ "attn.add_q_proj",
39
+ "attn.to_k",
40
+ "attn.add_v_proj",
41
+ "attn.to_v",
42
+ "attn.to_q"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": null,
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }