| | import json |
| | import glob |
| | import sys |
| | from collections import defaultdict |
| | from statistics import mean, variance |
| |
|
| | TARGET_LANGS = { |
| | "python", |
| | "java", |
| | "c/c++", |
| | "fortran", |
| | "r", |
| | "matlab", |
| | "shell", |
| | "rust", |
| | "go", |
| | } |
| |
|
| |
|
| | |
| | |
| | |
| | def load_jsonl_data(pattern="*.jsonl"): |
| | JSONL_FILES = glob.glob(pattern) |
| | |
| | |
| |
|
| | language_count = defaultdict(int) |
| | field_data = defaultdict(list) |
| | field_data_by_lang = defaultdict(lambda: defaultdict(list)) |
| |
|
| | for filename in JSONL_FILES: |
| | with open(filename, "r", encoding="utf-8") as f: |
| | for line in f: |
| | line = line.strip() |
| | if not line: |
| | continue |
| |
|
| | obj = json.loads(line) |
| | lang = obj.get("language", "unknown") |
| | language_count[lang] += 1 |
| |
|
| | |
| | for k, v in obj.items(): |
| | if isinstance(v, (int, float)): |
| | field_data[k].append(v) |
| | field_data_by_lang[lang][k].append(v) |
| |
|
| | return language_count, field_data, field_data_by_lang |
| |
|
| |
|
| | |
| | |
| | |
| | def print_language_distribution(language_count): |
| | print("\n========== 语言分布(language counts & percentage) ==========") |
| | total_items = sum(language_count.values()) |
| |
|
| | for lang, count in sorted(language_count.items(), key=lambda x: -x[1]): |
| | pct = count / total_items * 100 |
| | print(f"{lang}: {count} ({pct:.2f}%)") |
| |
|
| |
|
| | |
| | |
| | |
| | def compute_bins(start, end, step): |
| | bins = list(range(start, end + step, step)) |
| | labels = [f"{bins[i]}-{bins[i+1]}" for i in range(len(bins) - 1)] |
| | labels.append(f"{end}+") |
| | return bins, labels |
| |
|
| |
|
| | def compute_distribution(values, bins, labels): |
| | dist = {label: 0 for label in labels} |
| |
|
| | for v in values: |
| | placed = False |
| | for i in range(len(bins) - 1): |
| | if bins[i] <= v < bins[i + 1]: |
| | dist[labels[i]] += 1 |
| | placed = True |
| | break |
| | if not placed: |
| | dist[labels[-1]] += 1 |
| | return dist |
| |
|
| |
|
| | def print_distribution(title, dist, total_count): |
| | print(f"{title}") |
| | for label, count in dist.items(): |
| | pct = count / total_count * 100 |
| | print(f" {label}: {count} ({pct:.2f}%)") |
| |
|
| |
|
| | |
| | |
| | |
| | def analyze_field_distribution(jsonl_dir, field, start, end, step): |
| | print(f"\n================= 分析字段:{field} =================") |
| |
|
| | |
| | language_count, field_data, field_data_by_lang = load_jsonl_data(jsonl_dir) |
| |
|
| | |
| | print_language_distribution(language_count) |
| |
|
| | |
| | if field not in field_data: |
| | print(f"\n字段 '{field}' 在数据中不存在!") |
| | return |
| |
|
| | values = [] |
| | for lang in TARGET_LANGS: |
| | values.extend(field_data_by_lang.get(lang, {}).get(field, [])) |
| |
|
| | print(f"\n========== {field} 整体统计 ==========") |
| | print(f"个数: {len(values)}") |
| | print(f"最小值: {min(values)}") |
| | print(f"最大值: {max(values)}") |
| | print(f"均值: {mean(values):.2f}") |
| | if len(values) >= 2: |
| | print(f"方差: {variance(values):.2f}") |
| | else: |
| | print("方差: N/A") |
| |
|
| | |
| | bins, labels = compute_bins(start, end, step) |
| |
|
| | |
| | overall_dist = compute_distribution(values, bins, labels) |
| | print_distribution(f"区间分布", overall_dist, len(values)) |
| |
|
| | |
| | print(f"\n========== 按语言统计 {field} ==========") |
| |
|
| | for lang in TARGET_LANGS: |
| | fields = field_data_by_lang.get(lang) |
| | if not fields or field not in fields: |
| | continue |
| |
|
| | vals = fields[field] |
| |
|
| | print(f"\n--- {lang} ---") |
| | print(f"数量: {len(vals)}") |
| | print(f"最小值: {min(vals)}") |
| | print(f"最大值: {max(vals)}") |
| | print(f"均值: {mean(vals):.2f}") |
| | if len(vals) >= 2: |
| | print(f"方差: {variance(vals):.2f}") |
| | else: |
| | print("方差: N/A") |
| |
|
| | |
| | dist = compute_distribution(vals, bins, labels) |
| | print_distribution("区间分布:", dist, len(vals)) |
| |
|
| |
|
| | |
| | if __name__ == "__main__": |
| | jsonl_dir = "/home/weifengsun/tangou1/domain_code/src/datasets/analysis2/*.jsonl" |
| | |
| | |
| | field = "comment_lines" |
| | start = 0 |
| | end = 200 |
| | step = 20 |
| | analyze_field_distribution(jsonl_dir, field, start, end, step) |
| |
|