Tunisian_reddit / comment_scraper.py
Lime1's picture
Synced repo using 'sync_with_huggingface' Github Action
7958a44
raw
history blame
1.95 kB
import praw
import pandas as pd
import os
from datetime import datetime
FILENAME_COMMENTS = 'output_comments_after2022.csv'
def create_reddit_instance():
reddit = praw.Reddit(
client_id=os.getenv('REDDIT_CLIENT_ID'),
client_secret=os.getenv('REDDIT_CLIENT_SECRET'),
user_agent=os.getenv('REDDIT_USER_AGENT')
)
return reddit
def get_subreddit(reddit, subreddit_name):
subreddit = reddit.subreddit(subreddit_name)
return subreddit
def load_existing_data(file_name):
if os.path.exists(file_name):
df = pd.read_csv(file_name)
existing_ids = df['id'].tolist()
else:
df = pd.DataFrame()
existing_ids = []
return df, existing_ids
def get_new_comment_row(comment):
date = datetime.fromtimestamp(comment.created)
new_row = {
"id": comment.id,
"url": "https://www.reddit.com" + comment.permalink,
"score": comment.score,
"body": comment.body,
"date": date
}
return new_row
def save_data(df, file_name):
df.to_csv(file_name, index=False)
def main():
reddit = create_reddit_instance()
subreddit = get_subreddit(reddit, 'tunisia')
df_comments, existing_comment_ids = load_existing_data(FILENAME_COMMENTS)
print('Starting to scrape comments')
# Fetch 1000 newest comments
new_comments = list(subreddit.comments(limit=1000))
for comment in new_comments:
if comment.id in existing_comment_ids or comment.created_utc < 1672444800: # Skip comments before 2022
print(f'Skipped comment {comment.id}')
continue
new_row = get_new_comment_row(comment)
df_comments = df_comments._append(new_row, ignore_index=True)
save_data(df_comments, FILENAME_COMMENTS)
print(f'Loaded comment {comment.id}')
print('Finished scraping')
print("Data saved to ", FILENAME_COMMENTS)
if __name__ == "__main__":
main()