|
import pandas as pd |
|
import datasets |
|
from datasets import Features, Value, ClassLabel, Image, DownloadManager |
|
import logging |
|
import requests |
|
from PIL.Image import Image as PIL_Image |
|
import io |
|
|
|
_URLS = { |
|
"my_data": "https://drive.google.com/uc?export=download&id=1t7qllYbonoCgNzzh7w9NhmnMZ4pmqERo", |
|
} |
|
|
|
class RedditPoliticalSubs(datasets.GeneratorBasedBuilder): |
|
"""A Dataset builder for a DataFrame with Reddit data.""" |
|
|
|
VERSION = datasets.Version('1.0.0') |
|
|
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
description=("This dataset contains Reddit posts with various attributes."), |
|
features=Features({ |
|
"author": Value("string"), |
|
"created_utc": Value("string"), |
|
"domain": Value("string"), |
|
"title": Value("string"), |
|
"selftext": Value("string"), |
|
"subreddit": Value("string"), |
|
"score": Value("int32"), |
|
"num_comments": Value("int32"), |
|
"ups": Value("float32"), |
|
"downs": Value("float32"), |
|
"permalink": Value("string"), |
|
"is_self": Value("bool"), |
|
"url": Value("string"), |
|
"subreddit_subscribers": Value("float32"), |
|
"upvote_ratio": Value("float32"), |
|
"is_original_content": Value("string"), |
|
"media": Value("string"), |
|
"selftext_html": Value("string"), |
|
"author_flair_text": Value("string"), |
|
"link_flair_text": Value("string"), |
|
"image": Image(), |
|
"image_text": Value("string"), |
|
}), |
|
supervised_keys=None, |
|
homepage='https://www.reddit.com/', |
|
citation="", |
|
) |
|
|
|
def _split_generators(self, dl_manager: DownloadManager): |
|
downloaded_file = dl_manager.download_and_extract(_URLS["my_data"]) |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs={"dataframe_path": downloaded_file}, |
|
), |
|
] |
|
|
|
def _generate_examples(self, dataframe_path): |
|
"""Yields examples.""" |
|
df = pd.read_csv(dataframe_path) |
|
|
|
for idx, row in df.iterrows(): |
|
|
|
yield idx, { |
|
"author": row["author"], |
|
"created_utc": row["created_utc"], |
|
"domain": row["domain"] if pd.notna(row["domain"]) else "", |
|
"title": row["title"], |
|
"selftext": row["selftext"] if pd.notna(row["selftext"]) else "", |
|
"subreddit": row["subreddit"], |
|
"score": row["score"], |
|
"num_comments": row["num_comments"], |
|
"ups": row["ups"] if pd.notna(row["ups"]) else 0, |
|
"downs": row["downs"] if pd.notna(row["downs"]) else 0, |
|
"permalink": row["permalink"], |
|
"is_self": row["is_self"], |
|
"url": row["url"] if pd.notna(row["url"]) else "", |
|
"subreddit_subscribers": row["subreddit_subscribers"] if pd.notna(row["subreddit_subscribers"]) else 0.0, |
|
"upvote_ratio": row["upvote_ratio"] if pd.notna(row["upvote_ratio"]) else 0.0, |
|
"is_original_content": row["is_original_content"] if pd.notna(row["is_original_content"]) else False, |
|
"media": row["media"] if pd.notna(row["media"]) else "", |
|
"selftext_html": row["selftext_html"] if pd.notna(row["selftext_html"]) else "", |
|
"author_flair_text": row["author_flair_text"] if pd.notna(row["author_flair_text"]) else "", |
|
"link_flair_text": row["link_flair_text"] if pd.notna(row["link_flair_text"]) else "", |
|
"image": row['url'] if pd.notna(row['url']) else "", |
|
"image_text": row['image_text'] if pd.notna(row['image_text']) else "", |
|
} |