|
--- |
|
license: mit |
|
language: |
|
- en |
|
pretty_name: X |
|
--- |
|
|
|
This dataset contains tweets related to the Israel-Palestine conflict from October 17, 2023, to December 17, 2023. It includes information on tweet IDs, links, text, date, likes, and comments, categorized into different ranges of like counts. |
|
|
|
## Dataset Details |
|
|
|
- **Date Range:** October 17, 2023 - December 17, 2023 |
|
- **Total Tweets:** 15,478 |
|
- **Unique Tweets:** 14,854 |
|
|
|
## Data Description |
|
|
|
The dataset consists of the following columns: |
|
|
|
| Column | Description | |
|
|------------|-----------------------------------------------------------| |
|
| `id` | Unique identifier for the tweet | |
|
| `link` | URL link to the tweet | |
|
| `text` | Text content of the tweet | |
|
| `date` | Date and time when the tweet was posted | |
|
| `likes` | Number of likes the tweet received | |
|
| `comments` | Number of comments the tweet received | |
|
| `Label` | Like count range category | |
|
| `Count` | Number of tweets in the like count range category | |
|
|
|
|
|
|
|
|
|
## How to Process the Data |
|
|
|
To process the dataset, you can use the following Python code. This code reads the CSV file, cleans the tweets, tokenizes and lemmatizes the text, and filters out non-English tweets. |
|
|
|
### Required Libraries |
|
|
|
Make sure you have the following libraries installed: |
|
|
|
```bash |
|
pip install pandas nltk langdetect |
|
``` |
|
|
|
## Data Processing Code |
|
|
|
Here’s the code to process the tweets: |
|
|
|
```python |
|
import pandas as pd |
|
import re |
|
from nltk.tokenize import word_tokenize |
|
from nltk.corpus import stopwords |
|
from nltk.stem import WordNetLemmatizer |
|
from langdetect import detect, LangDetectException |
|
|
|
# Define the TweetProcessor class |
|
class TweetProcessor: |
|
def __init__(self, file_path): |
|
""" |
|
Initialize the object with the path to the CSV file. |
|
""" |
|
self.df = pd.read_csv(file_path) |
|
# Convert 'text' column to string type |
|
self.df['text'] = self.df['text'].astype(str) |
|
|
|
def clean_tweet(self, tweet): |
|
""" |
|
Clean a tweet by removing links, special characters, and extra spaces. |
|
""" |
|
# Remove links |
|
tweet = re.sub(r'https\S+', '', tweet, flags=re.MULTILINE) |
|
# Remove special characters and numbers |
|
tweet = re.sub(r'\W', ' ', tweet) |
|
# Replace multiple spaces with a single space |
|
tweet = re.sub(r'\s+', ' ', tweet) |
|
# Remove leading and trailing spaces |
|
tweet = tweet.strip() |
|
return tweet |
|
|
|
def tokenize_and_lemmatize(self, tweet): |
|
""" |
|
Tokenize and lemmatize a tweet by converting to lowercase, removing stopwords, and lemmatizing. |
|
""" |
|
# Tokenize the text |
|
tokens = word_tokenize(tweet) |
|
# Remove punctuation and numbers, and convert to lowercase |
|
tokens = [word.lower() for word in tokens if word.isalpha()] |
|
# Remove stopwords |
|
stop_words = set(stopwords.words('english')) |
|
tokens = [word for word in tokens if word not in stop_words] |
|
# Lemmatize the tokens |
|
lemmatizer = WordNetLemmatizer() |
|
tokens = [lemmatizer.lemmatize(word) for word in tokens] |
|
# Join tokens back into a single string |
|
return ' '.join(tokens) |
|
|
|
def process_tweets(self): |
|
""" |
|
Apply cleaning and lemmatization functions to the tweets in the DataFrame. |
|
""" |
|
def lang(x): |
|
try: |
|
return detect(x) == 'en' |
|
except LangDetectException: |
|
return False |
|
|
|
# Filter tweets for English language |
|
self.df = self.df[self.df['text'].apply(lang)] |
|
|
|
# Apply cleaning function |
|
self.df['cleaned_text'] = self.df['text'].apply(self.clean_tweet) |
|
# Apply tokenization and lemmatization function |
|
self.df['tokenized_and_lemmatized'] = self.df['cleaned_text'].apply(self.tokenize_and_lemmatize) |
|
|
|
``` |
|
|
|
Feel free to add or modify any details according to your specific requirements! |
|
|
|
Let me know if there’s anything else you’d like to adjust or add! |
|
|
|
|
|
|
|
|
|
|
|
|
|
## Usage |
|
|
|
This dataset can be used for various research purposes, including sentiment analysis, trend analysis, and event impact studies related to the Israel-Palestine conflict. |
|
For questions or feedback, please contact: |
|
|
|
- **Name:** Mehyar Mlaweh |
|
- **Email:** mehyarmlaweh0@gmail.com |