Josephgflowers's picture
Upload find-reason-fine.py
073b267 verified
raw
history blame
8.29 kB
import pandas as pd
import re
from concurrent.futures import ProcessPoolExecutor
from tqdm import tqdm
import os
import glob
# Science keywords (formatted for regex word boundaries)
science_keywords_list = [
# Core Types of Reasoning
"deductive reasoning", "inductive reasoning", "abductive reasoning",
"deductive logic", "inductive logic", "probabilistic reasoning",
"hypothetical reasoning", "falsifiability", "meta-cognition",
# Logic Structures and Components
"syllogism", "proposition", "premise", "conclusion", "logical fallacy",
"argument", "logical consistency", "logical operator", "step by step",
# Analytical and Critical Thinking
"critical thinking", "analytical skills", "creative thinking",
"convergent thinking", "divergent thinking", "contextual analysis",
"pattern recognition", "structured reflection", "reasoned judgment",
"cognitive load", "counterfactual thinking", "comparative reasoning",
"subjective reasoning", "objective reasoning", "systematic approach",
# Hypothesis and Evidence Analysis
"hypothesis testing", "hypothesis generation", "evidence-based reasoning",
"empirical reasoning", "evidence synthesis", "confirmation bias",
"cognitive bias", "causation vs correlation", "construct validity",
# Problem Solving and Decision Making
"problem analysis", "brainstorming", "decision making", "decision fatigue",
"satisficing", "bounded rationality", "opportunity cost",
"cost-benefit analysis", "optimization", "strategic planning",
"trade-off analysis", "prioritization matrix", "value prioritization",
# Heuristics and Algorithms
"heuristic", "heuristic reasoning", "algorithm", "recursive thinking",
"pattern matching", "dynamic programming", "systematic approach",
# Data Analysis and Modeling
"data analysis", "causal reasoning", "correlation", "probabilistic inference",
"qualitative analysis", "quantitative analysis", "predictive modeling",
"belief revision", "mental modeling", "proportional reasoning",
# Cognitive Processes and Biases
"cognitive dissonance", "framing effect", "bias mitigation",
"normative reasoning", "intuitive thinking", "belief bias",
# Argumentation and Discourse
"counterargument", "debate", "dialectic", "socratic questioning",
"disjunctive reasoning", "conjunctive reasoning", "chain of thought",
# Problem Decomposition and Structuring
"root cause analysis", "5 whys", "decision tree", "flow chart",
"process mapping", "mind mapping", "ideation", "brainwriting",
"problem decomposition", "value chain analysis",
# Analytical Frameworks and Techniques
"SWOT analysis", "gap analysis", "risk assessment", "scenario planning",
"simulation", "backcasting", "game theory", "decision matrix",
"opportunity analysis", "knowledge representation",
# Creative Thinking and Synthesis
"lateral thinking", "synergistic thinking", "brainstorming",
"synthesis", "ideation", "hypothetical deduction",
# Additional Analytical Techniques
"comparative analysis", "analytical hierarchy process", "multicriteria decision analysis",
"trade-off analysis", "constraint analysis", "thought experiment",
]
# Escape special regex characters and add word boundaries
science_keywords = [
r"\b" + re.escape(keyword).replace(r'\ ', ' ') + r"\b" for keyword in science_keywords_list
]
# Combine science keywords into a single regex pattern using non-capturing groups
science_regex = r'(?:' + r'|'.join(science_keywords) + r')'
# Function to process a chunk of the dataset
def process_chunk(chunk):
# Assign column names if they are not already set
if list(chunk.columns) != ['score', 'text', 'url']:
chunk.columns = ['score', 'text', 'url']
# Use vectorized string operations for efficiency
# Count the number of matches in each column
score_counts = chunk['score'].astype(str).str.count(science_regex, flags=re.IGNORECASE)
url_counts = chunk['url'].astype(str).str.count(science_regex, flags=re.IGNORECASE)
text_counts = chunk['text'].astype(str).str.count(science_regex, flags=re.IGNORECASE)
# Handle NaN values by filling them with zero
score_counts = score_counts.fillna(0)
url_counts = url_counts.fillna(0)
text_counts = text_counts.fillna(0)
# Sum the counts to get the science score
match_counts = score_counts + url_counts + text_counts
match_counts = match_counts.astype(int)
#
#
# Set a threshold for the minimum science score
threshold = 15 # Adjust this value as needed
#
#
# Filter rows that meet the threshold
filtered_chunk = chunk[match_counts >= threshold].copy()
filtered_chunk['science_score'] = match_counts[match_counts >= threshold]
# Replace the original 'score' with 'science_score'
filtered_chunk['score'] = filtered_chunk['science_score']
filtered_chunk = filtered_chunk.drop(columns=['science_score'])
return filtered_chunk
# Function to process a single CSV file
def process_file(input_file, output_file):
# Read the CSV file in chunks, assuming no header in the CSV file
chunk_size = 10000 # Adjust this value based on your memory constraints
reader = pd.read_csv(input_file, chunksize=chunk_size, header=None)
# Prepare the output file
first_chunk = True
# Number of worker processes
num_workers = 8 # Adjust based on your CPU cores
# Batch size for chunks to process in parallel
batch_size = num_workers * 4 # Adjust based on memory constraints
chunk_list = []
with ProcessPoolExecutor(max_workers=num_workers) as executor:
for chunk in tqdm(reader, desc=f'Reading chunks from {os.path.basename(input_file)}'):
chunk_list.append(chunk)
if len(chunk_list) == batch_size:
# Process batch of chunks in parallel
futures = [executor.submit(process_chunk, c) for c in chunk_list]
for future in tqdm(futures, desc='Processing batch', leave=False):
filtered_chunk = future.result()
if not filtered_chunk.empty:
if first_chunk:
filtered_chunk.to_csv(output_file, mode='w', index=False, header=False)
first_chunk = False
else:
filtered_chunk.to_csv(output_file, mode='a', index=False, header=False)
chunk_list = []
# Process any remaining chunks
if chunk_list:
futures = [executor.submit(process_chunk, c) for c in chunk_list]
for future in tqdm(futures, desc='Processing last batch', leave=False):
filtered_chunk = future.result()
if not filtered_chunk.empty:
if first_chunk:
filtered_chunk.to_csv(output_file, mode='w', index=False, header=False)
first_chunk = False
else:
filtered_chunk.to_csv(output_file, mode='a', index=False, header=False)
print(f'Finished processing {input_file}')
# List of directories to process
data_dir = '/media/joe/512-3/csv'
years = [f'CC-MAIN-{year}' for year in range(2013, 2025)] # Adjust years as needed
directories = [os.path.join(data_dir, year) for year in years]
# Process each CSV file in each directory
for dir_path in directories:
if not os.path.isdir(dir_path):
print(f'Directory not found: {dir_path}')
continue
csv_files = glob.glob(os.path.join(dir_path, '*.csv'))
print(f'Found {len(csv_files)} CSV files in {dir_path}')
for input_file in csv_files:
# Construct output file name
base_name = os.path.basename(input_file)
output_file = os.path.join(
dir_path, 'reason_' + base_name
)
# Check if output file already exists to avoid reprocessing
if os.path.exists(output_file):
print(f'Output file already exists. Skipping: {output_file}')
continue
process_file(input_file, output_file)