reddit_political_subs / reddit_political_subs.py
mo-mittal's picture
Update reddit_political_subs.py
e2bf408 verified
raw
history blame
No virus
3.94 kB
import pandas as pd
import datasets
from datasets import Features, Value, ClassLabel, Image, DownloadManager
import logging
import requests
from PIL.Image import Image as PIL_Image
import io
_URLS = {
"my_data": "https://drive.google.com/uc?export=download&id=1t7qllYbonoCgNzzh7w9NhmnMZ4pmqERo",
}
class RedditPoliticalSubs(datasets.GeneratorBasedBuilder):
"""A Dataset builder for a DataFrame with Reddit data."""
VERSION = datasets.Version('1.0.0')
def _info(self):
return datasets.DatasetInfo(
description=("This dataset contains Reddit posts with various attributes."),
features=Features({
"author": Value("string"),
"created_utc": Value("string"),
"domain": Value("string"),
"title": Value("string"),
"selftext": Value("string"),
"subreddit": Value("string"),
"score": Value("int32"),
"num_comments": Value("int32"),
"ups": Value("float32"),
"downs": Value("float32"),
"permalink": Value("string"),
"is_self": Value("bool"),
"url": Value("string"),
"subreddit_subscribers": Value("float32"),
"upvote_ratio": Value("float32"),
"is_original_content": Value("string"),
"media": Value("string"),
"selftext_html": Value("string"),
"author_flair_text": Value("string"),
"link_flair_text": Value("string"),
"image": Image(),
"image_text": Value("string"),
}),
supervised_keys=None,
homepage='https://www.reddit.com/',
citation="",
)
def _split_generators(self, dl_manager: DownloadManager):
downloaded_file = dl_manager.download_and_extract(_URLS["my_data"])
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={"dataframe_path": downloaded_file}, # light config
),
]
def _generate_examples(self, dataframe_path): #light config
"""Yields examples."""
df = pd.read_csv(dataframe_path)
for idx, row in df.iterrows():
yield idx, {
"author": row["author"],
"created_utc": row["created_utc"],
"domain": row["domain"] if pd.notna(row["domain"]) else "",
"title": row["title"],
"selftext": row["selftext"] if pd.notna(row["selftext"]) else "",
"subreddit": row["subreddit"],
"score": row["score"],
"num_comments": row["num_comments"],
"ups": row["ups"] if pd.notna(row["ups"]) else 0,
"downs": row["downs"] if pd.notna(row["downs"]) else 0,
"permalink": row["permalink"],
"is_self": row["is_self"],
"url": row["url"] if pd.notna(row["url"]) else "",
"subreddit_subscribers": row["subreddit_subscribers"] if pd.notna(row["subreddit_subscribers"]) else 0.0,
"upvote_ratio": row["upvote_ratio"] if pd.notna(row["upvote_ratio"]) else 0.0,
"is_original_content": row["is_original_content"] if pd.notna(row["is_original_content"]) else False,
"media": row["media"] if pd.notna(row["media"]) else "",
"selftext_html": row["selftext_html"] if pd.notna(row["selftext_html"]) else "",
"author_flair_text": row["author_flair_text"] if pd.notna(row["author_flair_text"]) else "",
"link_flair_text": row["link_flair_text"] if pd.notna(row["link_flair_text"]) else "",
"image": row['url'] if pd.notna(row['url']) else "",
"image_text": row['image_text'] if pd.notna(row['image_text']) else "",
}