| | """ |
| | Statistics for UD Vietnamese Dataset (UDD-v0.1) |
| | """ |
| |
|
| | from collections import Counter |
| | from os.path import dirname, join |
| |
|
| |
|
| | def parse_conllu(filepath): |
| | """Parse CoNLL-U file and return sentences.""" |
| | sentences = [] |
| | current_sentence = { |
| | 'tokens': [], |
| | 'upos': [], |
| | 'deprel': [], |
| | 'head': [], |
| | 'metadata': {} |
| | } |
| |
|
| | with open(filepath, 'r', encoding='utf-8') as f: |
| | for line in f: |
| | line = line.strip() |
| | if not line: |
| | if current_sentence['tokens']: |
| | sentences.append(current_sentence) |
| | current_sentence = { |
| | 'tokens': [], |
| | 'upos': [], |
| | 'deprel': [], |
| | 'head': [], |
| | 'metadata': {} |
| | } |
| | elif line.startswith('#'): |
| | |
| | if '=' in line: |
| | key, value = line[2:].split('=', 1) |
| | current_sentence['metadata'][key.strip()] = value.strip() |
| | else: |
| | parts = line.split('\t') |
| | if len(parts) >= 10: |
| | |
| | if '-' in parts[0] or '.' in parts[0]: |
| | continue |
| | current_sentence['tokens'].append(parts[1]) |
| | current_sentence['upos'].append(parts[3]) |
| | current_sentence['head'].append(parts[6]) |
| | current_sentence['deprel'].append(parts[7]) |
| |
|
| | |
| | if current_sentence['tokens']: |
| | sentences.append(current_sentence) |
| |
|
| | return sentences |
| |
|
| |
|
| | def compute_statistics(sentences): |
| | """Compute statistics from parsed sentences.""" |
| | stats = {} |
| |
|
| | |
| | stats['num_sentences'] = len(sentences) |
| | stats['num_tokens'] = sum(len(s['tokens']) for s in sentences) |
| |
|
| | |
| | sent_lengths = [len(s['tokens']) for s in sentences] |
| | stats['avg_sent_length'] = sum(sent_lengths) / len(sent_lengths) if sent_lengths else 0 |
| | stats['min_sent_length'] = min(sent_lengths) if sent_lengths else 0 |
| | stats['max_sent_length'] = max(sent_lengths) if sent_lengths else 0 |
| |
|
| | |
| | all_upos = [] |
| | for s in sentences: |
| | all_upos.extend(s['upos']) |
| | stats['upos_counts'] = Counter(all_upos) |
| |
|
| | |
| | all_deprel = [] |
| | for s in sentences: |
| | all_deprel.extend(s['deprel']) |
| | stats['deprel_counts'] = Counter(all_deprel) |
| |
|
| | |
| | depths = [] |
| | for s in sentences: |
| | max_depth = compute_tree_depth(s['head']) |
| | depths.append(max_depth) |
| | stats['avg_tree_depth'] = sum(depths) / len(depths) if depths else 0 |
| | stats['max_tree_depth'] = max(depths) if depths else 0 |
| |
|
| | |
| | root_upos = [] |
| | for s in sentences: |
| | for i, (upos, deprel) in enumerate(zip(s['upos'], s['deprel'])): |
| | if deprel == 'root': |
| | root_upos.append(upos) |
| | stats['root_upos_counts'] = Counter(root_upos) |
| |
|
| | return stats |
| |
|
| |
|
| | def compute_tree_depth(heads): |
| | """Compute maximum depth of dependency tree.""" |
| | n = len(heads) |
| | if n == 0: |
| | return 0 |
| |
|
| | depths = [0] * n |
| |
|
| | def get_depth(idx): |
| | if depths[idx] > 0: |
| | return depths[idx] |
| | head = int(heads[idx]) |
| | if head == 0: |
| | depths[idx] = 1 |
| | else: |
| | depths[idx] = get_depth(head - 1) + 1 |
| | return depths[idx] |
| |
|
| | for i in range(n): |
| | try: |
| | get_depth(i) |
| | except (RecursionError, IndexError): |
| | depths[i] = 1 |
| |
|
| | return max(depths) if depths else 0 |
| |
|
| |
|
| | def print_statistics(stats): |
| | """Print statistics in a nice format.""" |
| | print("=" * 60) |
| | print("UD Vietnamese Dataset (UDD-v0.1) Statistics") |
| | print("=" * 60) |
| |
|
| | print("\n## Basic Statistics") |
| | print(f" Sentences: {stats['num_sentences']:,}") |
| | print(f" Tokens: {stats['num_tokens']:,}") |
| | print(f" Avg sent length: {stats['avg_sent_length']:.2f}") |
| | print(f" Min sent length: {stats['min_sent_length']}") |
| | print(f" Max sent length: {stats['max_sent_length']}") |
| | print(f" Avg tree depth: {stats['avg_tree_depth']:.2f}") |
| | print(f" Max tree depth: {stats['max_tree_depth']}") |
| |
|
| | print("\n## UPOS Distribution") |
| | print(f" {'Tag':<10} {'Count':>8} {'Percent':>8}") |
| | print(" " + "-" * 28) |
| | total_tokens = stats['num_tokens'] |
| | for tag, count in stats['upos_counts'].most_common(): |
| | pct = count / total_tokens * 100 |
| | print(f" {tag:<10} {count:>8,} {pct:>7.2f}%") |
| |
|
| | print("\n## DEPREL Distribution") |
| | print(f" {'Relation':<20} {'Count':>8} {'Percent':>8}") |
| | print(" " + "-" * 38) |
| | for rel, count in stats['deprel_counts'].most_common(): |
| | pct = count / total_tokens * 100 |
| | print(f" {rel:<20} {count:>8,} {pct:>7.2f}%") |
| |
|
| | print("\n## Root UPOS Distribution") |
| | print(f" {'UPOS':<10} {'Count':>8} {'Percent':>8}") |
| | print(" " + "-" * 28) |
| | total_roots = sum(stats['root_upos_counts'].values()) |
| | for tag, count in stats['root_upos_counts'].most_common(): |
| | pct = count / total_roots * 100 |
| | print(f" {tag:<10} {count:>8,} {pct:>7.2f}%") |
| |
|
| | print("\n" + "=" * 60) |
| |
|
| |
|
| | def main(): |
| | |
| | base_dir = dirname(dirname(__file__)) |
| | conllu_file = join(base_dir, 'train.conllu') |
| |
|
| | print(f"Reading: {conllu_file}") |
| | sentences = parse_conllu(conllu_file) |
| |
|
| | stats = compute_statistics(sentences) |
| | print_statistics(stats) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|