File size: 1,428 Bytes
d4f9d26 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 |
'''
This is an example of how the data was collected for a large subreddit (post level).
'''
import pandas as pd
import json
master_df = pd.DataFrame()
path = '/content/drive/MyDrive/project 1 data (sta 663)/The_Donald/'
chunks = []
for i in range(1,9):
chunks.append(path + 'The_Donald_submissions.00' + str(i))
# chunks = ["/content/drive/MyDrive/project 1 data (sta 663)/The_Donald"]
# read line by line
for chunk in chunks:
data = []
with open(chunk, 'r') as file:
for line in file:
try:
# Parse the JSON line and append to the list
data.append(json.loads(line))
except json.JSONDecodeError as e:
# Output the error and skip this line
print(f"Error decoding JSON: {e}") # usually due to post deletion or incorrect type
df = pd.DataFrame(data)
# 'created_utc' to a datetime column
df['created_utc'] = pd.to_datetime(df['created_utc'], unit='s')
# extract the year from the 'created_utc' column
df['year'] = df['created_utc'].dt.year
# filtering the years for 2014 to the current year
df_filtered = df[(df['year'] >= 2014)]
master_df = pd.concat([master_df, df_filtered])
# can sort and get the top 100 posts per year by score
master_df_sorted = master_df.sort_values(by=['year', 'score'], ascending=[True, False])
top_posts_per_year = master_df_sorted.groupby('year').head(100) |