| | """ |
| | Visualization module: Generate publication-ready figures (PNG/SVG) |
| | ACL conference style, 1920x1080 or A4 landscape size |
| | """ |
| | import matplotlib |
| | matplotlib.use('Agg') |
| | import matplotlib.pyplot as plt |
| | import matplotlib.font_manager as fm |
| | import seaborn as sns |
| | import pandas as pd |
| | import numpy as np |
| | from pathlib import Path |
| | import json |
| | import statistics |
| | from collections import Counter |
| |
|
| | |
| | |
| | font_families_to_try = ['Arial', 'DejaVu Sans', 'Liberation Sans', 'sans-serif'] |
| | available_fonts = [f.name for f in fm.fontManager.ttflist] |
| | font_found = None |
| |
|
| | for font_family in font_families_to_try: |
| | |
| | font_lower = font_family.lower() |
| | if any(f.lower() == font_lower for f in available_fonts): |
| | font_found = font_family |
| | break |
| |
|
| | if font_found is None: |
| | |
| | font_found = 'sans-serif' |
| |
|
| | |
| | |
| | |
| | plt.rcParams['font.family'] = font_found |
| | plt.rcParams['font.size'] = 24 |
| | plt.rcParams['axes.labelsize'] = 42 |
| | plt.rcParams['axes.titlesize'] = 30 |
| | plt.rcParams['xtick.labelsize'] = 36 |
| | plt.rcParams['ytick.labelsize'] = 36 |
| | plt.rcParams['legend.fontsize'] = 20 |
| | plt.rcParams['figure.titlesize'] = 32 |
| | plt.rcParams['axes.linewidth'] = 1.5 |
| | plt.rcParams['axes.spines.top'] = False |
| | plt.rcParams['axes.spines.right'] = False |
| | plt.rcParams['axes.grid'] = True |
| | plt.rcParams['grid.alpha'] = 0.3 |
| | plt.rcParams['grid.linewidth'] = 0.5 |
| | plt.rcParams['axes.unicode_minus'] = False |
| |
|
| | |
| | NATURE_COLORS = { |
| | 'primary': '#2E5090', |
| | 'secondary': '#1A5490', |
| | 'accent': '#4A90E2', |
| | 'success': '#2E7D32', |
| | 'warning': '#F57C00', |
| | 'error': '#C62828', |
| | 'neutral': '#424242', |
| | 'light': '#E3F2FD' |
| | } |
| |
|
| | |
| | nature_palette = ['#2E5090', '#4A90E2', '#1A5490', '#6BA3D8', '#94C4E8'] |
| |
|
| |
|
| | class Visualization: |
| | def __init__(self, output_dir, figsize=(19.2, 10.8), dpi=150): |
| | self.output_dir = Path(output_dir) |
| | self.output_dir.mkdir(parents=True, exist_ok=True) |
| | self.figsize = figsize |
| | self.dpi = dpi |
| | self.fig_counter = 1 |
| | |
| | def apply_nature_style(self, ax): |
| | """Apply Nature journal style to axes""" |
| | ax.spines['top'].set_visible(False) |
| | ax.spines['right'].set_visible(False) |
| | ax.spines['left'].set_linewidth(1.5) |
| | ax.spines['bottom'].set_linewidth(1.5) |
| | ax.grid(True, alpha=0.3, linestyle='--', linewidth=0.5) |
| | ax.tick_params(width=1.5, length=5) |
| | |
| | def save_fig(self, fig, name): |
| | """保存图片""" |
| | fig_path_png = self.output_dir / f"fig_{self.fig_counter:02d}_{name}.png" |
| | fig_path_svg = self.output_dir / f"fig_{self.fig_counter:02d}_{name}.svg" |
| | fig.savefig(fig_path_png, dpi=self.dpi, bbox_inches='tight', facecolor='white') |
| | fig.savefig(fig_path_svg, bbox_inches='tight', facecolor='white') |
| | plt.close(fig) |
| | self.fig_counter += 1 |
| | print(f"Saved: {fig_path_png}") |
| | |
| | def plot_funnel(self, stage_a_dir, stage_b_dir, repo_meta_dir, top_n=None): |
| | """绘制漏斗图:搜索->过滤->深度分析""" |
| | |
| | stage_a_path = Path(stage_a_dir) / 'summary_overall.json' |
| | stage_b_path = Path(stage_b_dir) / 'filter_summary.json' |
| | repo_meta_path = Path(repo_meta_dir) / 'repo_meta_summary.json' |
| | |
| | with open(stage_a_path, 'r') as f: |
| | stage_a = json.load(f) |
| | |
| | with open(stage_b_path, 'r') as f: |
| | stage_b = json.load(f) |
| | |
| | |
| | deep_analysis_count = top_n if top_n else 0 |
| | try: |
| | with open(repo_meta_path, 'r') as f: |
| | repo_meta = json.load(f) |
| | deep_analysis_count = repo_meta.get('total_repos', deep_analysis_count) |
| | except: |
| | pass |
| | |
| | |
| | if top_n: |
| | deep_analysis_label = f'Deep Analysis\n({top_n:,} repos)' |
| | else: |
| | deep_analysis_label = f'Deep Analysis\n({deep_analysis_count:,} repos)' |
| | |
| | stages = ['Search Stage\n(~1.3M repos)', 'Filtered\n(~30K repos)', deep_analysis_label] |
| | values = [ |
| | stage_a['total_records'], |
| | stage_b['total'], |
| | deep_analysis_count |
| | ] |
| | |
| | fig, ax = plt.subplots(figsize=self.figsize) |
| | self.apply_nature_style(ax) |
| | |
| | |
| | y_pos = np.arange(len(stages)) |
| | colors = [NATURE_COLORS['primary'], NATURE_COLORS['accent'], NATURE_COLORS['secondary']] |
| | |
| | bars = ax.barh(y_pos, values, color=colors, alpha=0.85, edgecolor='white', linewidth=2) |
| | ax.set_yticks(y_pos) |
| | ax.set_yticklabels(stages, fontsize=36) |
| | ax.set_xlabel('Number of Repositories', fontsize=42, fontweight='bold') |
| | ax.set_title('Data Pipeline: Search → Filter → Deep Analysis', fontsize=30, fontweight='bold', pad=20) |
| | |
| | |
| | for i, (bar, val) in enumerate(zip(bars, values)): |
| | ax.text(val * 0.5, i, f'{val:,}', ha='center', va='center', |
| | fontsize=24, fontweight='bold', color='white') |
| | |
| | ax.invert_yaxis() |
| | plt.tight_layout() |
| | |
| | self.save_fig(fig, 'funnel') |
| | |
| | def plot_top_keywords(self, csv_path, top_n=20): |
| | """绘制Top keywords条形图""" |
| | df = pd.read_csv(csv_path) |
| | df_top = df.head(top_n) |
| | |
| | fig, ax = plt.subplots(figsize=self.figsize) |
| | self.apply_nature_style(ax) |
| | |
| | |
| | bars = ax.barh(range(len(df_top)), df_top['count'], |
| | color=NATURE_COLORS['primary'], alpha=0.85, |
| | edgecolor='white', linewidth=1.5) |
| | ax.set_yticks(range(len(df_top))) |
| | ax.set_yticklabels(df_top['keyword'], fontsize=36, rotation=15, ha='right') |
| | ax.set_xlabel('Number of Repositories', fontsize=42, fontweight='bold') |
| | ax.set_title(f'Top {top_n} Keywords (Search Stage)', fontsize=30, fontweight='bold', pad=20) |
| | |
| | |
| | for i, (idx, row) in enumerate(df_top.iterrows()): |
| | ax.text(row['count'] * 0.5, i, f"{int(row['count']):,}", |
| | ha='center', va='center', fontsize=24, fontweight='bold', color='white') |
| | |
| | ax.invert_yaxis() |
| | plt.tight_layout() |
| | |
| | self.save_fig(fig, 'top_keywords') |
| | |
| | def plot_language_distribution(self, csv_path, top_n=15): |
| | """绘制语言分布""" |
| | df = pd.read_csv(csv_path) |
| | df = df[df['language'] != '<empty>'].head(top_n) |
| | |
| | fig, ax = plt.subplots(figsize=self.figsize) |
| | self.apply_nature_style(ax) |
| | |
| | bars = ax.bar(range(len(df)), df['count'], |
| | color=NATURE_COLORS['primary'], alpha=0.85, |
| | edgecolor='white', linewidth=1.5) |
| | ax.set_xticks(range(len(df))) |
| | ax.set_xticklabels(df['language'], rotation=45, ha='right', fontsize=36) |
| | ax.set_ylabel('Number of Repositories', fontsize=42, fontweight='bold') |
| | ax.set_title(f'Top {top_n} Programming Languages (Search Stage)', |
| | fontsize=30, fontweight='bold', pad=20) |
| | |
| | |
| | for i, count in enumerate(df['count']): |
| | ax.text(i, count, f"{int(count):,}", ha='center', va='bottom', |
| | fontsize=24, fontweight='bold', rotation=45) |
| | |
| | plt.tight_layout() |
| | |
| | self.save_fig(fig, 'language_distribution') |
| | |
| | def plot_stars_distribution(self, csv_path): |
| | """绘制stars分布(对数坐标)""" |
| | df = pd.read_csv(csv_path) |
| | |
| | fig, ax = plt.subplots(figsize=self.figsize) |
| | self.apply_nature_style(ax) |
| | |
| | |
| | log_bins = np.logspace(0, np.log10(df['stars'].max() + 1), 50) |
| | ax.hist(df['stars'], bins=log_bins, |
| | color=NATURE_COLORS['primary'], alpha=0.75, |
| | edgecolor='white', linewidth=1) |
| | ax.set_xscale('log') |
| | ax.set_xlabel('Stars (log scale)', fontsize=42, fontweight='bold') |
| | ax.set_ylabel('Number of Repositories', fontsize=42, fontweight='bold') |
| | ax.set_title('Distribution of Repository Stars (Log Scale)', |
| | fontsize=30, fontweight='bold', pad=20) |
| | |
| | plt.tight_layout() |
| | |
| | self.save_fig(fig, 'stars_distribution') |
| | |
| | def plot_filter_results(self, csv_path): |
| | """绘制过滤结果(按keyword的YES/NO)""" |
| | df = pd.read_csv(csv_path) |
| | df = df.head(15) |
| | |
| | fig, ax = plt.subplots(figsize=self.figsize) |
| | self.apply_nature_style(ax) |
| | |
| | x = np.arange(len(df)) |
| | width = 0.4 |
| | |
| | bars1 = ax.bar(x - width/2, df['yes'], width, label='Relevant (YES)', |
| | color=NATURE_COLORS['success'], alpha=0.85, edgecolor='white', linewidth=1.5) |
| | bars2 = ax.bar(x + width/2, df['no'], width, label='Irrelevant (NO)', |
| | color=NATURE_COLORS['error'], alpha=0.85, edgecolor='white', linewidth=1.5) |
| | |
| | ax.set_xlabel('Keyword', fontsize=42, fontweight='bold') |
| | ax.set_ylabel('Number of Repositories', fontsize=42, fontweight='bold') |
| | ax.set_title('Filter Results: YES/NO Distribution by Keyword', |
| | fontsize=30, fontweight='bold', pad=20) |
| | ax.set_xticks(x) |
| | ax.set_xticklabels(df['keyword'], rotation=45, ha='right', fontsize=36) |
| | ax.legend(fontsize=20, frameon=True, fancybox=True, shadow=True) |
| | |
| | plt.tight_layout() |
| | |
| | self.save_fig(fig, 'filter_results_by_keyword') |
| | |
| | def plot_reason_length_comparison(self, csv_path): |
| | """绘制reason长度对比(YES vs NO)""" |
| | df = pd.read_csv(csv_path) |
| | |
| | fig, ax = plt.subplots(figsize=self.figsize) |
| | self.apply_nature_style(ax) |
| | |
| | yes_lengths = df[df['label'] == 'YES']['length'] |
| | no_lengths = df[df['label'] == 'NO']['length'] |
| | |
| | bp = ax.boxplot([yes_lengths, no_lengths], labels=['YES', 'NO'], |
| | patch_artist=True, |
| | widths=0.6, |
| | boxprops=dict(facecolor=NATURE_COLORS['primary'], alpha=0.7, linewidth=2), |
| | medianprops=dict(color='white', linewidth=3), |
| | whiskerprops=dict(linewidth=2), |
| | capprops=dict(linewidth=2), |
| | flierprops=dict(marker='o', markersize=8, alpha=0.5)) |
| | |
| | ax.set_ylabel('Reason Length (Characters)', fontsize=42, fontweight='bold') |
| | ax.set_title('Comparison of Filter Reason Length: YES vs NO', |
| | fontsize=30, fontweight='bold', pad=20) |
| | ax.set_xticklabels(['Relevant (YES)', 'Irrelevant (NO)'], fontsize=36) |
| | |
| | plt.tight_layout() |
| | |
| | self.save_fig(fig, 'reason_length_comparison') |
| | |
| | def plot_extension_distribution(self, csv_path, top_n=20, repo_top_n=None): |
| | """绘制文件扩展名分布""" |
| | df = pd.read_csv(csv_path) |
| | df = df.head(top_n) |
| | |
| | fig, ax = plt.subplots(figsize=self.figsize) |
| | self.apply_nature_style(ax) |
| | |
| | bars = ax.barh(range(len(df)), df['count'], |
| | color=NATURE_COLORS['accent'], alpha=0.85, |
| | edgecolor='white', linewidth=1.5) |
| | ax.set_yticks(range(len(df))) |
| | ax.set_yticklabels(df['extension'], fontsize=36, rotation=15, ha='right') |
| | ax.set_xlabel('Number of Files', fontsize=42, fontweight='bold') |
| | |
| | |
| | if repo_top_n: |
| | repo_label = f'Top {repo_top_n:,} Repositories' |
| | else: |
| | repo_label = 'All Repositories' |
| | ax.set_title(f'Top {top_n} File Extension Distribution ({repo_label})', |
| | fontsize=30, fontweight='bold', pad=20) |
| | |
| | ax.invert_yaxis() |
| | plt.tight_layout() |
| | |
| | self.save_fig(fig, 'extension_distribution') |
| | |
| | def plot_repo_file_count_distribution(self, csv_path, repo_top_n=None): |
| | """绘制仓库文件数分布""" |
| | df = pd.read_csv(csv_path) |
| | |
| | fig, ax = plt.subplots(figsize=self.figsize) |
| | self.apply_nature_style(ax) |
| | |
| | |
| | log_bins = np.logspace(0, np.log10(df['total_files'].max() + 1), 50) |
| | ax.hist(df['total_files'], bins=log_bins, |
| | color=NATURE_COLORS['primary'], alpha=0.75, |
| | edgecolor='white', linewidth=1) |
| | ax.set_xscale('log') |
| | ax.set_xlabel('Number of Files (log scale)', fontsize=42, fontweight='bold') |
| | ax.set_ylabel('Number of Repositories', fontsize=42, fontweight='bold') |
| | |
| | |
| | if repo_top_n: |
| | repo_label = f'Top {repo_top_n:,} Repositories' |
| | else: |
| | repo_label = 'All Repositories' |
| | ax.set_title(f'Distribution of Repository File Counts ({repo_label})', |
| | fontsize=30, fontweight='bold', pad=20) |
| | |
| | plt.tight_layout() |
| | |
| | self.save_fig(fig, 'repo_file_count_distribution') |
| | |
| | def plot_stars_vs_code_size(self, repos_searched_csv, repo_level_csv, repo_top_n=None): |
| | """绘制stars vs 代码规模散点图(需要join)""" |
| | |
| | df_searched = pd.read_csv(repos_searched_csv, usecols=['full_name', 'stars']) |
| | df_searched = df_searched.dropna(subset=['stars']) |
| | df_searched['stars'] = df_searched['stars'].astype(float) |
| | |
| | |
| | df_repo = pd.read_csv(repo_level_csv) |
| | df_repo['full_name'] = df_repo['full_name'].fillna(df_repo['repo_name'].str.replace('___', '/')) |
| | |
| | |
| | df_merged = df_repo.merge(df_searched, on='full_name', how='inner') |
| | df_merged = df_merged[df_merged['total_code_lines'] > 0] |
| | |
| | if len(df_merged) == 0: |
| | print("Warning: No data to plot stars vs code size") |
| | return |
| | |
| | fig, ax = plt.subplots(figsize=self.figsize) |
| | self.apply_nature_style(ax) |
| | |
| | |
| | ax.scatter(df_merged['total_code_lines'], df_merged['stars'], |
| | alpha=0.4, s=30, color=NATURE_COLORS['primary'], edgecolors='white', linewidth=0.5) |
| | ax.set_xscale('log') |
| | ax.set_yscale('log') |
| | ax.set_xlabel('Lines of Code (LOC, log scale)', fontsize=42, fontweight='bold') |
| | ax.set_ylabel('Stars (log scale)', fontsize=42, fontweight='bold') |
| | |
| | |
| | if repo_top_n: |
| | repo_label = f'Top {repo_top_n:,} Repositories' |
| | else: |
| | repo_label = 'All Repositories' |
| | ax.set_title(f'Stars vs Code Size ({repo_label})', |
| | fontsize=30, fontweight='bold', pad=20) |
| | |
| | |
| | corr = np.corrcoef(np.log10(df_merged['total_code_lines']), |
| | np.log10(df_merged['stars']))[0, 1] |
| | ax.text(0.05, 0.95, f'Correlation: r = {corr:.3f}', transform=ax.transAxes, |
| | fontsize=24, verticalalignment='top', fontweight='bold', |
| | bbox=dict(boxstyle='round', facecolor='white', alpha=0.8, edgecolor=NATURE_COLORS['primary'], linewidth=2)) |
| | |
| | plt.tight_layout() |
| | |
| | self.save_fig(fig, 'stars_vs_code_size') |
| | |
| | def plot_repo_stats_by_language(self, repo_level_csv, repo_top_n=None): |
| | """绘制按主语言的代码统计对比""" |
| | df = pd.read_csv(repo_level_csv) |
| | df = df[df['primary_language'] != 'unknown'] |
| | |
| | |
| | top_langs = df['primary_language'].value_counts().head(10).index |
| | df = df[df['primary_language'].isin(top_langs)] |
| | |
| | |
| | larger_figsize = (24, 16) |
| | fig, axes = plt.subplots(2, 2, figsize=larger_figsize) |
| | |
| | colors_list = [NATURE_COLORS['primary'], NATURE_COLORS['accent'], |
| | NATURE_COLORS['success'], NATURE_COLORS['secondary']] |
| | |
| | |
| | ax = axes[0, 0] |
| | self.apply_nature_style(ax) |
| | lang_stats = df.groupby('primary_language')['total_code_lines'].mean().sort_values(ascending=False) |
| | lang_stats.plot(kind='bar', ax=ax, color=colors_list[0], alpha=0.85, edgecolor='white', linewidth=1.5) |
| | ax.set_title('Average Lines of Code', fontsize=28, fontweight='bold', pad=25) |
| | ax.set_xlabel('') |
| | ax.set_ylabel('Average LOC', fontsize=42) |
| | ax.tick_params(axis='x', rotation=45, labelsize=36, pad=10) |
| | ax.tick_params(axis='y', labelsize=36, pad=10) |
| | |
| | |
| | ax = axes[0, 1] |
| | self.apply_nature_style(ax) |
| | lang_stats = df.groupby('primary_language')['total_functions'].mean().sort_values(ascending=False) |
| | lang_stats.plot(kind='bar', ax=ax, color=colors_list[1], alpha=0.85, edgecolor='white', linewidth=1.5) |
| | ax.set_title('Average Number of Functions', fontsize=28, fontweight='bold', pad=25) |
| | ax.set_xlabel('') |
| | ax.set_ylabel('Average Functions', fontsize=42) |
| | ax.tick_params(axis='x', rotation=45, labelsize=36, pad=10) |
| | ax.tick_params(axis='y', labelsize=36, pad=10) |
| | |
| | |
| | ax = axes[1, 0] |
| | self.apply_nature_style(ax) |
| | lang_stats = df.groupby('primary_language')['comment_ratio'].mean().sort_values(ascending=False) |
| | lang_stats.plot(kind='bar', ax=ax, color=colors_list[2], alpha=0.85, edgecolor='white', linewidth=1.5) |
| | ax.set_title('Average Comment Ratio', fontsize=28, fontweight='bold', pad=25) |
| | ax.set_xlabel('') |
| | ax.set_ylabel('Comment Ratio', fontsize=42) |
| | ax.tick_params(axis='x', rotation=45, labelsize=36, pad=10) |
| | ax.tick_params(axis='y', labelsize=36, pad=10) |
| | |
| | |
| | ax = axes[1, 1] |
| | self.apply_nature_style(ax) |
| | lang_stats = df.groupby('primary_language')['language_entropy'].mean().sort_values(ascending=False) |
| | lang_stats.plot(kind='bar', ax=ax, color=colors_list[3], alpha=0.85, edgecolor='white', linewidth=1.5) |
| | ax.set_title('Average Language Diversity', fontsize=28, fontweight='bold', pad=25) |
| | ax.set_xlabel('') |
| | ax.set_ylabel('Language Entropy', fontsize=42) |
| | ax.tick_params(axis='x', rotation=45, labelsize=36, pad=10) |
| | ax.tick_params(axis='y', labelsize=36, pad=10) |
| | |
| | |
| | if repo_top_n: |
| | repo_label = f'Top {repo_top_n:,} Repositories' |
| | else: |
| | repo_label = 'All Repositories' |
| | plt.suptitle(f'Repository Code Statistics by Primary Language ({repo_label})', |
| | fontsize=36, fontweight='bold', y=0.995) |
| | |
| | |
| | |
| | plt.subplots_adjust(hspace=0.8, wspace=0.4, top=0.88, bottom=0.08, left=0.08, right=0.95) |
| | |
| | self.save_fig(plt.gcf(), 'repo_stats_by_language') |
| | |
| | def plot_keyword_wordcloud(self, csv_path, max_words=200): |
| | """Generate wordcloud for keywords (Nature style: colorful)""" |
| | try: |
| | from wordcloud import WordCloud |
| | except ImportError: |
| | print("Warning: wordcloud library not installed. Skipping wordcloud generation.") |
| | print("Install with: pip install wordcloud") |
| | return |
| | |
| | try: |
| | |
| | df = pd.read_csv(csv_path) |
| | |
| | |
| | keyword_freq = {} |
| | for keyword, count in zip(df['keyword'], df['count']): |
| | keyword_freq[str(keyword)] = int(count) |
| | |
| | |
| | wordcloud = WordCloud( |
| | width=1920, |
| | height=1080, |
| | background_color='white', |
| | colormap='Blues', |
| | max_words=max_words, |
| | relative_scaling=0.5, |
| | min_font_size=10, |
| | prefer_horizontal=0.7, |
| | collocations=False |
| | ) |
| | |
| | |
| | wordcloud.generate_from_frequencies(keyword_freq) |
| | |
| | |
| | fig, ax = plt.subplots(figsize=self.figsize) |
| | |
| | |
| | |
| | pil_image = wordcloud.to_image() |
| | wordcloud_array = np.array(pil_image) |
| | |
| | ax.imshow(wordcloud_array, interpolation='bilinear') |
| | ax.axis('off') |
| | ax.set_title('Keyword Word Cloud (Search Stage)', |
| | fontsize=30, fontweight='bold', pad=20) |
| | |
| | plt.tight_layout() |
| | |
| | |
| | fig_path_png = self.output_dir / f"fig_{self.fig_counter:02d}_keyword_wordcloud.png" |
| | fig.savefig(fig_path_png, dpi=self.dpi, bbox_inches='tight', facecolor='white') |
| | plt.close(fig) |
| | self.fig_counter += 1 |
| | print(f"Saved: {fig_path_png}") |
| | |
| | except Exception as e: |
| | print(f"Error generating wordcloud: {e}") |
| | import traceback |
| | traceback.print_exc() |
| |
|
| |
|
| | def generate_all_visualizations(stage_a_dir, stage_b_dir, repo_meta_dir, code_stats_dir, repos_searched_csv, top_n=None): |
| | """生成所有图表""" |
| | viz = Visualization(Path(stage_a_dir).parent / 'figures') |
| | |
| | print("Generating visualizations...") |
| | |
| | |
| | top_n_suffix = f"_top{top_n}" if top_n else "" |
| | |
| | |
| | try: |
| | viz.plot_funnel(stage_a_dir, stage_b_dir, repo_meta_dir, top_n=top_n) |
| | except Exception as e: |
| | print(f"Error plotting funnel: {e}") |
| | |
| | try: |
| | viz.plot_top_keywords(Path(stage_a_dir) / 'by_keyword.csv') |
| | except Exception as e: |
| | print(f"Error plotting top keywords: {e}") |
| | |
| | try: |
| | viz.plot_keyword_wordcloud(Path(stage_a_dir) / 'by_keyword.csv') |
| | except Exception as e: |
| | print(f"Error generating keyword wordcloud: {e}") |
| | |
| | try: |
| | viz.plot_language_distribution(Path(stage_a_dir) / 'by_language.csv') |
| | except Exception as e: |
| | print(f"Error plotting language distribution: {e}") |
| | |
| | try: |
| | viz.plot_stars_distribution(Path(stage_a_dir) / 'stars_distribution.csv') |
| | except Exception as e: |
| | print(f"Error plotting stars distribution: {e}") |
| | |
| | |
| | try: |
| | viz.plot_filter_results(Path(stage_b_dir) / 'filter_by_keyword.csv') |
| | except Exception as e: |
| | print(f"Error plotting filter results: {e}") |
| | |
| | try: |
| | viz.plot_reason_length_comparison(Path(stage_b_dir) / 'reason_length_distribution.csv') |
| | except Exception as e: |
| | print(f"Error plotting reason length comparison: {e}") |
| | |
| | |
| | try: |
| | viz.plot_extension_distribution(Path(repo_meta_dir) / 'extension_distribution.csv', repo_top_n=top_n) |
| | except Exception as e: |
| | print(f"Error plotting extension distribution: {e}") |
| | |
| | try: |
| | viz.plot_repo_file_count_distribution(Path(repo_meta_dir) / f'repo_meta_scan{top_n_suffix}.csv', repo_top_n=top_n) |
| | except Exception as e: |
| | print(f"Error plotting repo file count distribution: {e}") |
| | |
| | |
| | try: |
| | viz.plot_stars_vs_code_size(repos_searched_csv, Path(code_stats_dir) / f'repo_level_metrics{top_n_suffix}.csv', repo_top_n=top_n) |
| | except Exception as e: |
| | print(f"Error plotting stars vs code size: {e}") |
| | |
| | try: |
| | viz.plot_repo_stats_by_language(Path(code_stats_dir) / f'repo_level_metrics{top_n_suffix}.csv', repo_top_n=top_n) |
| | except Exception as e: |
| | print(f"Error plotting repo stats by language: {e}") |
| | |
| | print(f"Visualization complete! All figures saved to {viz.output_dir}") |
| |
|
| |
|