sriram882004 commited on
Commit
92eb4c6
·
verified ·
1 Parent(s): 7ffd0e4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -7
README.md CHANGED
@@ -1,26 +1,64 @@
1
  ---
2
  license: mit
3
  ---
 
4
  # SQL Socratic Models
5
 
6
- This repository contains fine-tuned LLMs for Text-to-SQL using Socratic prompting and reasoning decomposition.
7
 
8
  ## Models
9
  - phi3_rq4
10
  - qwen25
11
- - llama31
12
 
13
  ## Method
14
- We fine-tune models on SQL generation tasks with Socratic reasoning prompts.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
  ## Contributions
17
- - Fine-tuning across multiple architectures
18
- - Ablation studies on reasoning components
19
- - Evaluation on Text-to-SQL benchmarks
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  ## Usage
22
  ```python
23
  from transformers import AutoModelForCausalLM, AutoTokenizer
24
 
25
  model = AutoModelForCausalLM.from_pretrained("sriram882004/SQL-Socratic-Models/phi3_rq4")
26
- tokenizer = AutoTokenizer.from_pretrained(...)
 
1
  ---
2
  license: mit
3
  ---
4
+
5
  # SQL Socratic Models
6
 
7
+ This repository contains fine-tuned large language models for **Socratic SQL instruction** in higher education, focusing on guiding learners through SQL concepts using structured reasoning rather than providing direct solutions.
8
 
9
  ## Models
10
  - phi3_rq4
11
  - qwen25
12
+ - gemma2
13
 
14
  ## Method
15
+
16
+ Our approach is designed to support **conceptual learning in STEM education** through Socratic interaction:
17
+
18
+ - **Phase 1 (Data Construction):**
19
+ SQL instruction data is augmented with Socratic prompts emphasizing:
20
+ - Question decomposition
21
+ - Conceptual hints
22
+ - Guided reasoning steps
23
+
24
+ - **Phase 2 (Fine-Tuning):**
25
+ We apply full fine-tuning (FFT) on small, open-source LLMs with **pedagogical constraints** that explicitly discourage direct answer generation and instead promote:
26
+ - Conceptual scaffolding
27
+ - Incremental reasoning
28
+ - Learner-centered guidance
29
+
30
+ - **Phase 3 (Evaluation):**
31
+ Models are evaluated using:
32
+ - **BERTScore** for semantic alignment with expected reasoning
33
+ - **ROUGE-L** to measure and control **answer leakage** (i.e., avoidance of direct SQL solutions)
34
 
35
  ## Contributions
36
+ - Fine-tuning across multiple architectures (Phi-3, Qwen2.5, Gemma2) for **instructional SQL reasoning**
37
+ - Development of **Socratic SQL prompting framework** for higher education contexts
38
+ - Evaluation of models on their ability to generate **guidance without revealing final answers**
39
+ - Ablation study identifying factors that enable LLMs to mimic effective instructors through:
40
+ - Misconception-aware feedback
41
+ - Iterative questioning
42
+ - Structured reasoning support
43
+
44
+ ## Task
45
+
46
+ Given a natural language SQL question, the model generates:
47
+
48
+ 1. Socratic reasoning steps
49
+ 2. Conceptual hints and guiding questions
50
+ 3. Intermediate decomposition of the problem
51
+
52
+ **The model does NOT produce the final SQL query**, ensuring alignment with instructional use in higher education settings.
53
+
54
+ This design supports:
55
+ - Active learning
56
+ - Conceptual understanding of SQL
57
+ - Integration of database concepts into broader STEM curricula
58
 
59
  ## Usage
60
  ```python
61
  from transformers import AutoModelForCausalLM, AutoTokenizer
62
 
63
  model = AutoModelForCausalLM.from_pretrained("sriram882004/SQL-Socratic-Models/phi3_rq4")
64
+ tokenizer = AutoTokenizer.from_pretrained("sriram882004/SQL-Socratic-Models/phi3_rq4")