| | """ |
| | 修复版训练数据生成器 |
| | 核心改进: |
| | 1. 直接基于代码内容生成准确的问答对 |
| | 2. 不依赖LLM生成(避免循环依赖) |
| | 3. 使用模板化方法确保数据质量 |
| | 4. 优化项目概览问题,使其更具项目特色 |
| | """ |
| | import json |
| | import yaml |
| | import random |
| | from pathlib import Path |
| | from typing import List, Dict, Any |
| | from dataclasses import dataclass, field |
| | import re |
| | from collections import defaultdict |
| |
|
| |
|
| | @dataclass |
| | class TrainingSample: |
| | """训练样本""" |
| | conversations: List[Dict[str, str]] |
| | metadata: Dict[str, Any] |
| |
|
| |
|
| | class FixedDataGenerator: |
| | """修复版数据生成器 - 基于规则和模板""" |
| | |
| | def __init__(self, config_path: str = "../config/default_config.yaml", |
| | analysis_path: str = "../data/repository_analysis.json"): |
| | with open(config_path, 'r', encoding='utf-8') as f: |
| | self.config = yaml.safe_load(f) |
| | |
| | try: |
| | with open(analysis_path, 'r', encoding='utf-8') as f: |
| | self.analysis_data = json.load(f) |
| | except FileNotFoundError: |
| | print(f"❌ 警告: 找不到分析文件 {analysis_path}。请先运行分析器。") |
| | self.analysis_data = {'code_elements': [], 'project_context': {}} |
| |
|
| | self.code_elements = self.analysis_data.get('code_elements', []) |
| | self.project_context = self.analysis_data.get('project_context', {}) |
| | self.project_name = self.project_context.get('project_name', 'Laddr') |
| | |
| | self.training_samples = [] |
| | |
| | def generate_training_data(self): |
| | """生成训练数据""" |
| | print(f"Generating training data for {self.project_name}...") |
| | |
| | |
| | print("Generating code explanation samples...") |
| | self._generate_code_explanation_samples() |
| | |
| | |
| | print("Generating API usage samples...") |
| | self._generate_api_usage_samples() |
| | |
| | |
| | print("Generating project overview samples...") |
| | self._generate_project_overview_samples() |
| | |
| | |
| | print("Generating code location samples...") |
| | self._generate_code_location_samples() |
| | |
| | print(f"Total samples generated: {len(self.training_samples)}") |
| | |
| | def _generate_code_explanation_samples(self): |
| | """生成代码解释样本 - 基于真实代码和docstring""" |
| | |
| | candidates = [e for e in self.code_elements |
| | if e.get('docstring') and len(e.get('code', '')) > 50] |
| | |
| | for element in candidates[:300]: |
| | name = element['name'] |
| | docstring = element['docstring'] |
| | filepath = element['filepath'] |
| | element_type = element['type'] |
| | code = element.get('code', '') |
| | |
| | |
| | signature = self._extract_signature(code, element_type) |
| | |
| | |
| | questions = [ |
| | f"请解释 {self.project_name} 中 `{name}` 的作用。", |
| | f"{self.project_name} 的 `{name}` 是做什么的?", |
| | f"在 {self.project_name} 项目中,`{name}` 有什么功能?", |
| | ] |
| | question = random.choice(questions) |
| | |
| | |
| | answer_parts = [] |
| | |
| | |
| | answer_parts.append(f"`{name}` 是 {self.project_name} 项目中的一个 {self._type_to_cn(element_type)},位于 `{filepath}`。") |
| | |
| | |
| | if docstring: |
| | |
| | clean_doc = self._clean_docstring(docstring) |
| | answer_parts.append(f"\n**功能描述**:\n{clean_doc}") |
| | |
| | |
| | if signature: |
| | answer_parts.append(f"\n**函数签名**:\n```python\n{signature}\n```") |
| | |
| | |
| | params = element.get('parameters', []) |
| | if params and len(params) > 0: |
| | param_desc = "\n**参数**:\n" |
| | for param in params[:5]: |
| | param_name = param.get('name', 'unknown') |
| | param_type = param.get('type', 'Any') |
| | |
| | param_desc_from_doc = self._extract_param_desc(docstring, param_name) |
| | if param_desc_from_doc: |
| | param_info = f"- `{param_name}` ({param_type}): {param_desc_from_doc}\n" |
| | else: |
| | param_info = f"- `{param_name}` ({param_type})\n" |
| |
|
| | param_desc += param_info |
| | answer_parts.append(param_desc) |
| | |
| | |
| | return_type = element.get('return_type') |
| | if return_type: |
| | answer_parts.append(f"\n**返回值**:`{return_type}`") |
| | |
| | answer = ''.join(answer_parts) |
| | |
| | self.training_samples.append(TrainingSample( |
| | conversations=[ |
| | {"role": "user", "content": question}, |
| | {"role": "assistant", "content": answer} |
| | ], |
| | metadata={ |
| | "task_type": "code_explanation", |
| | "element_name": name, |
| | "filepath": filepath |
| | } |
| | )) |
| | |
| | def _generate_api_usage_samples(self): |
| | """生成API使用示例 - 基于函数签名""" |
| | |
| | candidates = [e for e in self.code_elements |
| | if e['type'] in ['function', 'method'] |
| | and not e['name'].startswith('_') |
| | and e.get('parameters')] |
| | |
| | for element in candidates[:150]: |
| | name = element['name'] |
| | params = element.get('parameters', []) |
| | filepath = element['filepath'] |
| | docstring = element.get('docstring', '') |
| | |
| | question = f"如何在 {self.project_name} 中使用 `{name}` 函数?" |
| | |
| | |
| | answer_parts = [] |
| | answer_parts.append(f"`{name}` 位于 `{filepath}`,使用方法如下:") |
| | |
| | |
| | param_names = [p['name'] for p in params if p['name'] != 'self'] |
| | if param_names: |
| | example_code = f"{name}(" |
| | param_examples = [] |
| | for p in param_names[:5]: |
| | param_examples.append(f"{p}=...") |
| | example_code += ", ".join(param_examples) |
| | example_code += ")" |
| | |
| | answer_parts.append(f"\n```python\n{example_code}\n```") |
| | |
| | |
| | if params: |
| | answer_parts.append("\n**参数说明**:") |
| | for param in params[:5]: |
| | if param['name'] != 'self': |
| | param_type = param.get('type', 'Any') |
| | |
| | param_desc_from_doc = self._extract_param_desc(docstring, param['name']) |
| | |
| | answer_parts.append(f"\n- `{param['name']}`: {param_type}") |
| | if param_desc_from_doc: |
| | answer_parts[-1] += f" - {param_desc_from_doc}" |
| | |
| | |
| | if docstring: |
| | clean_doc = self._clean_docstring(docstring)[:200] |
| | if clean_doc: |
| | answer_parts.append(f"\n\n**功能简述**:{clean_doc}...") |
| | |
| | answer = ''.join(answer_parts) |
| | |
| | self.training_samples.append(TrainingSample( |
| | conversations=[ |
| | {"role": "user", "content": question}, |
| | {"role": "assistant", "content": answer} |
| | ], |
| | metadata={ |
| | "task_type": "api_usage", |
| | "element_name": name |
| | } |
| | )) |
| | |
| | def _generate_project_overview_samples(self): |
| | """生成项目概览问答 - 基于统计信息""" |
| | stats = self.analysis_data.get('statistics', {}) |
| | description = self.project_context.get('description', '') |
| | techs = self.project_context.get('main_technologies', []) |
| | file_type_counts = self.analysis_data.get('statistics', {}).get('file_type_counts', {}) |
| | |
| | |
| | q1_list = [ |
| | f"请用一句话描述 {self.project_name} 项目的主要功能。", |
| | f"{self.project_name} 是一个什么样的项目?", |
| | f"简单介绍一下 {self.project_name} 项目。" |
| | ] |
| | q1 = random.choice(q1_list) |
| | |
| | a1_parts = [ |
| | f"{self.project_name} 是一个 Python 项目。" |
| | ] |
| | |
| | if description: |
| | |
| | a1_parts.append(f"\n**核心目标**:\n{description}") |
| | else: |
| | a1_parts.append("\n**核心目标**:此项目旨在提供一个可扩展的多代理系统框架(Agent Framework),支持任务规划、工具调用、消息队列和数据库集成等功能。") |
| | |
| | |
| | if techs: |
| | a1_parts.append(f"\n\n**主要技术栈**:{', '.join(techs[:5])}等。") |
| | |
| | a1_parts.append(f"\n\n项目包含 {stats.get('total_elements', 0)} 个代码元素,主要由 {stats.get('classes', 0)} 个类和 {stats.get('functions', 0) + stats.get('methods', 0)} 个函数/方法构成。") |
| | |
| | a1 = ''.join(a1_parts) |
| | |
| | self.training_samples.append(TrainingSample( |
| | conversations=[ |
| | {"role": "user", "content": q1}, |
| | {"role": "assistant", "content": a1} |
| | ], |
| | metadata={"task_type": "project_overview"} |
| | )) |
| | |
| | |
| | q2_list = [ |
| | f"{self.project_name} 的项目结构是怎样的?", |
| | f"请列举 {self.project_name} 的核心模块。", |
| | ] |
| | q2 = random.choice(q2_list) |
| | |
| | a2_parts = [f"{self.project_name} 项目包含以下主要部分:\n"] |
| | |
| | |
| | modules = self.project_context.get('key_modules', []) |
| | if modules: |
| | a2_parts.append("\n**核心模块**:\n") |
| | for mod in modules[:10]: |
| | a2_parts.append(f"- `{mod}`\n") |
| | else: |
| | a2_parts.append("\n**核心模块**:\n- `core` (核心逻辑,如Agent Runtime, Tooling, Config)\n- `cli` (命令行接口)\n- `llms` (LLM后端实现)\n") |
| |
|
| | |
| | if file_type_counts: |
| | file_stats = ', '.join(f'{k.lstrip(".").upper()}: {v}' for k, v in file_type_counts.items() if k not in ['.other']) |
| | a2_parts.append(f"\n**主要文件类型统计**:{file_stats}") |
| | |
| | a2 = ''.join(a2_parts) |
| | |
| | self.training_samples.append(TrainingSample( |
| | conversations=[ |
| | {"role": "user", "content": q2}, |
| | {"role": "assistant", "content": a2} |
| | ], |
| | metadata={"task_type": "project_structure"} |
| | )) |
| | |
| | |
| | top_elements = sorted(self.code_elements, |
| | key=lambda x: x.get('complexity', 0), |
| | reverse=True)[:10] |
| | |
| | q3 = f"{self.project_name} 中有哪些核心类和函数?" |
| | a3_parts = [f"{self.project_name} 的核心组件包括(基于复杂度和重要性):\n"] |
| | |
| | for elem in top_elements: |
| | name = elem['name'] |
| | filepath = elem['filepath'] |
| | elem_type = self._type_to_cn(elem['type']) |
| | |
| | doc = elem.get('docstring', '') |
| | short_doc = self._clean_docstring(doc).split('\n')[0][:80].strip() |
| | |
| | line = f"\n- `{name}` ({elem_type}):位于 `{filepath}`" |
| | if short_doc: |
| | line += f" - {short_doc}..." |
| | a3_parts.append(line) |
| | |
| | if len(top_elements) > 0: |
| | a3 = ''.join(a3_parts) |
| | self.training_samples.append(TrainingSample( |
| | conversations=[ |
| | {"role": "user", "content": q3}, |
| | {"role": "assistant", "content": a3} |
| | ], |
| | metadata={"task_type": "core_components"} |
| | )) |
| | |
| | def _generate_code_location_samples(self): |
| | """生成代码定位任务""" |
| | |
| | file_elements = defaultdict(list) |
| | for elem in self.code_elements: |
| | |
| | if elem['name'] == '__init__' and 'module' not in elem['type']: |
| | continue |
| | file_elements[elem['filepath']].append(elem) |
| | |
| | for filepath, elements in list(file_elements.items())[:50]: |
| | |
| | selected = random.sample(elements, min(3, len(elements))) |
| | |
| | for elem in selected: |
| | name = elem['name'] |
| | elem_type = self._type_to_cn(elem['type']) |
| | |
| | question = f"在 {self.project_name} 中,`{name}` {elem_type}在哪个文件里?" |
| | |
| | |
| | answer = f"`{name}` 位于 `{filepath}`。" |
| | |
| | self.training_samples.append(TrainingSample( |
| | conversations=[ |
| | {"role": "user", "content": question}, |
| | {"role": "assistant", "content": answer} |
| | ], |
| | metadata={ |
| | "task_type": "code_location", |
| | "element_name": name, |
| | "filepath": filepath |
| | } |
| | )) |
| | |
| | def _extract_signature(self, code: str, element_type: str) -> str: |
| | """提取函数/类签名""" |
| | if not code: |
| | return "" |
| | |
| | lines = code.strip().split('\n') |
| | signature_lines = [] |
| | |
| | for line in lines: |
| | line = line.strip() |
| | if not line: |
| | continue |
| | |
| | signature_lines.append(line) |
| | |
| | |
| | if element_type in ['function', 'method'] and (line.startswith('def ') or line.startswith('async def ')): |
| | |
| | if not line.endswith(':'): |
| | continue |
| | return '\n'.join(signature_lines) |
| | |
| | |
| | if element_type == 'class' and line.startswith('class '): |
| | if not line.endswith(':'): |
| | continue |
| | return '\n'.join(signature_lines) |
| |
|
| | |
| | if line.endswith((':')) and not line.startswith(('def ', 'class ')): |
| | break |
| |
|
| | |
| | return '\n'.join(signature_lines[:5]) |
| | |
| | def _clean_docstring(self, docstring: str) -> str: |
| | """清理docstring""" |
| | if not docstring: |
| | return "" |
| | |
| | |
| | lines = docstring.strip().split('\n') |
| | cleaned = [] |
| | for line in lines: |
| | line = line.strip() |
| | if line: |
| | cleaned.append(line) |
| | |
| | return ' '.join(cleaned) |
| |
|
| | def _extract_param_desc(self, docstring: str, param_name: str) -> str: |
| | """从 docstring 中尝试提取参数描述""" |
| | if not docstring: |
| | return "" |
| | |
| | match = re.search(rf"(?:Args|Parameters|Params):\s*(?:[\n\r]\s*-)?\s*`?{re.escape(param_name)}`?\s*[:\-]\s*(.*)", docstring, re.IGNORECASE) |
| | if match: |
| | desc = match.group(1).split('\n')[0].strip() |
| | return desc if desc else "无描述" |
| | return "" |
| | |
| | def _type_to_cn(self, element_type: str) -> str: |
| | """元素类型转中文""" |
| | mapping = { |
| | 'function': '函数', |
| | 'method': '方法', |
| | 'class': '类', |
| | 'variable': '变量', |
| | 'module': '模块' |
| | } |
| | return mapping.get(element_type, element_type) |
| | |
| | def save_training_data(self): |
| | """保存训练数据""" |
| | output_dir = Path(self.config['dataset']['output_dir']) |
| | output_dir.mkdir(parents=True, exist_ok=True) |
| | |
| | |
| | random.shuffle(self.training_samples) |
| | |
| | |
| | total = len(self.training_samples) |
| | train_size = int(total * 0.8) |
| | val_size = int(total * 0.1) |
| | |
| | if total < 10: |
| | train_size = max(1, total // 2) |
| | val_size = max(1, (total - train_size) // 2) |
| | |
| | |
| | if train_size + val_size > total: |
| | val_size = total - train_size |
| |
|
| | train_data = self.training_samples[:train_size] |
| | val_data = self.training_samples[train_size:train_size + val_size] |
| | test_data = self.training_samples[train_size + val_size:] |
| | |
| | |
| | self._save_jsonl(train_data, output_dir / "train.jsonl") |
| | self._save_jsonl(val_data, output_dir / "val.jsonl") |
| | self._save_jsonl(test_data, output_dir / "test.jsonl") |
| | |
| | |
| | metadata = { |
| | 'total_samples': total, |
| | 'train_samples': len(train_data), |
| | 'val_samples': len(val_data), |
| | 'test_samples': len(test_data), |
| | 'project_name': self.project_name, |
| | 'task_distribution': self._get_task_distribution() |
| | } |
| | |
| | with open(output_dir / "metadata.json", 'w', encoding='utf-8') as f: |
| | json.dump(metadata, f, indent=2, ensure_ascii=False) |
| | |
| | print(f"\n✓ Training data saved:") |
| | print(f" Train: {len(train_data)}") |
| | print(f" Val: {len(val_data)}") |
| | print(f" Test: {len(test_data)}") |
| | print(f" Total: {total}") |
| | |
| | |
| | print(f"\n📝 Sample training example:") |
| | if train_data: |
| | sample = random.choice(train_data) |
| | print(f"Q: {sample.conversations[0]['content'][:100]}...") |
| | print(f"A: {sample.conversations[1]['content'][:150]}...") |
| | |
| | def _save_jsonl(self, data: List[TrainingSample], filepath: Path): |
| | """保存为JSONL格式""" |
| | with open(filepath, 'w', encoding='utf-8') as f: |
| | for sample in data: |
| | |
| | json.dump({'conversations': sample.conversations}, f, ensure_ascii=False) |
| | f.write('\n') |
| | |
| | def _get_task_distribution(self) -> Dict[str, int]: |
| | """统计任务分布""" |
| | dist = {} |
| | for sample in self.training_samples: |
| | task_type = sample.metadata.get('task_type', 'unknown') |
| | dist[task_type] = dist.get(task_type, 0) + 1 |
| | return dist |
| |
|
| |
|
| | def main(): |
| | print("="*60) |
| | print("Fixed Training Data Generator (Project-Specific Answers Enhanced)") |
| | print("="*60) |
| | |
| | generator = FixedDataGenerator() |
| | generator.generate_training_data() |
| | generator.save_training_data() |
| | |
| | print("\n" + "="*60) |
| | print("✓ Data generation completed!") |
| | print("="*60) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|