| | import os |
| | import pandas as pd |
| | from datetime import timedelta,datetime |
| |
|
| | from lib.utilities import codebook |
| | from lib.experiment_specs import study_config |
| |
|
| |
|
| | from lib.data_helpers.builder_utils import BuilderUtils |
| | from lib.data_helpers import test |
| |
|
| | from lib.utilities import serialize |
| |
|
| | class Gaming(): |
| | gaming_dir = os.path.join("data","external","intermediate","PhoneDashboard","Gaming") |
| | events_file = os.path.join(gaming_dir,"Events") |
| | first_last_file = os.path.join(gaming_dir, "FirstLast") |
| | diagnosed_file = os.path.join(gaming_dir, "Diagnosed") |
| | diagnosed_test_file = diagnosed_file.replace("intermediate","intermediate_test") |
| |
|
| | good_diag = ["Phone never shut off", |
| | "Phone shut off", |
| | "Phone shut off, even if d1<d0", |
| | ] |
| |
|
| | game_cols = ["AppCode", "AppRuntime", "ForegroundApp","BatteryLevel", "DeviceRuntime", "CreatedDatetime","CreatedEasternDatetime", "ScreenActive", |
| | "CreatedDate", "Sequence","Zipfile","TimeZone"] |
| |
|
| |
|
| | """ - detects suspicious events of a granular dataframe that comes from an individual zipfile |
| | - records the first and last pings in each granular data frame to scan first and last """ |
| | @staticmethod |
| | def scan(df,file,first_last_bool = False): |
| | for col in Gaming.game_cols: |
| | if col in df.columns: |
| | df[f"Prev{col}"] = df[col].shift(1) |
| | keep_cols = [x + y for x in ["", "Prev"] for y in Gaming.game_cols] |
| |
|
| | |
| | if first_last_bool == False: |
| | ev = df.loc[(df["PrevAppRuntime"]>df["AppRuntime"]) & (df["PrevAppCode"]==df["AppCode"]),keep_cols] |
| |
|
| | else: |
| | |
| | |
| | ev = df.loc[(df["PrevAppRuntime"]>df["AppRuntime"]) |
| | & (df["PrevAppCode"]==df["AppCode"]) |
| | & (df["Sequence"]=="First") |
| | & (df["PrevSequence"]=="Last") |
| | & (df["PrevZipfile"]!=df["Zipfile"]) |
| | , keep_cols] |
| |
|
| | serialize.save_pickle(ev, os.path.join(Gaming.gaming_dir,"Granular",f"events{file}")) |
| |
|
| | """gets the first and last observation from each raw zipfile""" |
| | @staticmethod |
| | def get_first_last(df,file): |
| | first = df.groupby("AppCode").first() |
| | first["Sequence"] = "First" |
| | last = df.groupby("AppCode").last() |
| | last["Sequence"] = "Last" |
| | first_last_df = first.append(last).reset_index() |
| | first_last_df = first_last_df[Gaming.game_cols] |
| | serialize.save_pickle(first_last_df, os.path.join(Gaming.gaming_dir, "FirstLast", f"first_last_{file}")) |
| |
|
| | """ assembles the events file, diagnoses gaming events, and summarizes blackouts on the user level by phase""" |
| | @staticmethod |
| | def process_gaming(error_margin, hour_use,raw_user_df): |
| |
|
| | |
| | config_user_dict = serialize.open_yaml("config_user.yaml") |
| | if config_user_dict['local']['test']: |
| | diag_df = serialize.open_pickle(Gaming.diagnosed_file) |
| |
|
| | else: |
| | Gaming._add_first_last_events() |
| | ev_df = Gaming._aggregate_events() |
| | diag_df = Gaming._diagnose_events(ev_df, error_margin, hour_use) |
| |
|
| | |
| | game_user_df = Gaming._reshape_events(diag_df, raw_user_df) |
| |
|
| | |
| | game_user_df_SA = Gaming._reshape_events(diag_df.loc[diag_df["PrevScreenActive"]==1], raw_user_df,"ActiveBlackoutsOverPhase") |
| |
|
| | game_hour_df = Gaming._expand_gaming_df(diag_df,"GameHourDf") |
| |
|
| | |
| | |
| |
|
| | return game_user_df |
| |
|
| |
|
| | """ aggregate the first last observations, and then scan them. |
| | We are scanning if the last reading from the previous zipfile |
| | has a app runtime that is greater than the app runtime for the next zipfile """ |
| | @staticmethod |
| | def _add_first_last_events(): |
| | fl_dir = os.path.join(Gaming.gaming_dir, "FirstLast") |
| | df = pd.concat([serialize.soft_df_open(os.path.join(fl_dir, x)) for x in os.listdir(fl_dir)]) |
| | df = df.sort_values(by=["AppCode", "CreatedEasternDatetime"]).reset_index(drop=True) |
| | if datetime.now()>study_config.surveys["Baseline"]["Start"]: |
| | df = df.loc[df['CreatedDatetime']>study_config.surveys["Baseline"]["Start"]] |
| | df["PrevSequence"] = df["Sequence"].shift(1) |
| | Gaming.scan(df, "fl", first_last_bool=True) |
| |
|
| | """aggregates all the individual events in the granular directory""" |
| | @staticmethod |
| | def _aggregate_events(): |
| | ev_dir = os.path.join(Gaming.gaming_dir, "Granular") |
| | ev_df = pd.concat([serialize.soft_df_open(os.path.join(ev_dir, x)) for x in os.listdir(ev_dir)]) |
| | ev_df = ev_df.drop_duplicates(subset=["AppCode", "CreatedEasternDatetime"], keep='last').reset_index(drop=True) |
| | serialize.save_pickle(ev_df, Gaming.events_file) |
| | return ev_df |
| |
|
| |
|
| | """ estimates the runtime of the phone when the user was not tracking |
| | - d0: the device runtime right before pennyworth stopped recording |
| | - d1: the device runtime when PD returned to recording |
| | - dd: difference in phone runtime (d1 - d0) |
| | - td: difference in the timestamps associated with d0 and d1 |
| | - error_margin: number of hours that CreateDateTime or runtime stamps can deviate before error is flagged |
| | - all variables have hour units |
| | """ |
| |
|
| | @staticmethod |
| | def _diagnose_events(ev_df, error_margin, clean_hour_use): |
| | df = ev_df.sort_values(by = ['AppCode','CreatedEasternDatetime']) |
| | df = df.loc[df["PrevCreatedEasternDatetime"]>study_config.first_pull] |
| |
|
| | if datetime.now()>study_config.surveys["Baseline"]["Start"]: |
| | df = df.loc[df['PrevCreatedDatetime']>study_config.surveys["Baseline"]["Start"]] |
| |
|
| | df["CreatedEasternDatetimeDiffHours"] = (df["CreatedEasternDatetime"] - df["PrevCreatedEasternDatetime"]).apply( |
| | lambda x: round(x.days * 24 + x.seconds / (60 * 60), 2)) |
| |
|
| | for col in ["DeviceRuntime", "AppRuntime", "PrevDeviceRuntime", "PrevAppRuntime"]: |
| | df[f"{col}Hours"] = (df[f"{col}"] / (1000 * 60 * 60)).round(decimals=2) |
| |
|
| | for col in ["DeviceRuntimeHours","AppRuntimeHours"]: |
| | df[col+"Diff"] = df[col]-df[f"Prev{col}"] |
| |
|
| | ne_dict = df.to_dict(orient='index') |
| | day = clean_hour_use.groupby(["AppCode","CreatedDate"])["UseMinutes"].sum() |
| | day_dic = {k: day[k].to_dict() for k, v in day.groupby(level=0)} |
| |
|
| | for key, val in ne_dict.items(): |
| |
|
| | d0 = val["PrevDeviceRuntimeHours"] |
| | d1 = val["DeviceRuntimeHours"] |
| | td = val["CreatedEasternDatetimeDiffHours"] |
| | date0 = val["PrevCreatedDatetime"] |
| | date1 = val["CreatedDatetime"] |
| |
|
| | if val["AppCode"] in day_dic: |
| | app_dic = day_dic[val["AppCode"]] |
| | else: |
| | |
| | app_dic = {} |
| |
|
| | |
| | |
| | if (date0+timedelta(days=1)).date()<date1.date(): |
| | next_day = date0+timedelta(days=1) |
| | while next_day < date1.date(): |
| | if next_day in app_dic: |
| | ne_dict[key]['Diagnosis'] = "ERROR: FalsePositive" |
| | break |
| | else: |
| | next_day = next_day + timedelta(days=1) |
| |
|
| | if date1<date0: |
| | ne_dict[key]['Diagnosis'] = "ERROR: Date1<Date0" |
| |
|
| | |
| | elif d1 - d0 < 0: |
| | |
| | ne_dict[key]['Diagnosis'] = "Phone shut off" |
| | ne_dict[key]['BlackoutHoursLB'] = d1 |
| | ne_dict[key]['BlackoutHoursUB'] = td |
| |
|
| | if td + error_margin < d1: |
| | |
| | ne_dict[key]['Diagnosis'] = "ERROR: td <d1 | d1-d0 <= 0 " |
| |
|
| | else: |
| | if td >= d1: |
| | |
| | |
| | ne_dict[key]['Diagnosis'] = f"Phone shut off, even if d1<d0" |
| | ne_dict[key]['BlackoutHoursLB'] = d1 |
| | ne_dict[key]['BlackoutHoursUB'] = td |
| |
|
| | else: |
| | |
| | ne_dict[key]['Diagnosis'] = f"Phone never shut off" |
| | ne_dict[key]['BlackoutHoursLB'] = td |
| | ne_dict[key]['BlackoutHoursUB'] = td |
| |
|
| | if td + error_margin < d1 - d0: |
| | |
| | ne_dict[key]['Diagnosis'] = "ERROR: if phone never shutoff, no way for td < d1-d0" |
| |
|
| |
|
| |
|
| | df = pd.DataFrame.from_dict(ne_dict, orient='index') |
| | df["BlackoutHours"] = (df["BlackoutHoursLB"] + df["BlackoutHoursUB"])/2 |
| | df = Gaming._diagnose_dups(df) |
| | serialize.save_pickle(df, Gaming.diagnosed_file) |
| | test.save_test_df(df,Gaming.diagnosed_test_file) |
| |
|
| | return df |
| |
|
| | @staticmethod |
| | def _diagnose_dups(df): |
| | df = df.sort_values(by=["AppCode", "PrevCreatedEasternDatetime"]).reset_index(drop=True) |
| | d_dict = df.to_dict(orient='index') |
| | for key, val in d_dict.items(): |
| | if key + 1 not in d_dict: |
| | continue |
| |
|
| | if d_dict[key]["AppCode"] != d_dict[key + 1]["AppCode"]: |
| | continue |
| |
|
| | if d_dict[key]["CreatedEasternDatetime"] > d_dict[key + 1]["PrevCreatedEasternDatetime"]: |
| | d_dict[key]["Diagnosis"] = "Error: Another event starts before this event ends" |
| |
|
| | |
| | if d_dict[key]["CreatedEasternDatetime"] < d_dict[key + 1]["CreatedEasternDatetime"]: |
| | d_dict[key + 1]["Diagnosis"] = "Error: Another event ends after this event starts" |
| |
|
| | df = pd.DataFrame.from_dict(d_dict, orient='index') |
| | return df |
| |
|
| | """ |
| | Input: takes the diagnosed event level dataframe |
| | Output: User level df that contains the total blackout period time by phase |
| | """ |
| | @staticmethod |
| | def _reshape_events(diag_df,raw_user,file_name = None): |
| | df = diag_df.loc[diag_df["Diagnosis"].isin(Gaming.good_diag)] |
| | df["CreatedDate"] = df["CreatedDatetime"].apply(lambda x: x.date()) |
| | df = BuilderUtils.add_phase_label(df, |
| | raw_df_date = "CreatedDate", |
| | start_buffer = 0, |
| | end_buffer = -1,) |
| |
|
| | |
| | codes = [study_config.phases[x]["StartSurvey"]["Code"] for x in list(study_config.phases.keys())] |
| | rename_dic = dict(zip(list(study_config.phases.keys()), codes)) |
| | df["Phase"] = df["Phase"].apply(lambda x: rename_dic[x] if x in rename_dic else x) |
| |
|
| |
|
| | df_s = df.groupby(["AppCode","Phase"])["BlackoutHours"].sum().reset_index() |
| |
|
| | df_p = df_s.pivot_table(index=["AppCode"], |
| | values=["BlackoutHours"], |
| | columns=["Phase"], |
| | aggfunc='first') |
| |
|
| | |
| | df_p.columns = ['_'.join(col[::-1]).strip() for col in df_p.columns.values] |
| | df_p = df_p.reset_index() |
| |
|
| | |
| | |
| |
|
| | |
| | return df_p |
| |
|
| | @staticmethod |
| | def _expand_gaming_df(diag,file_name): |
| | ex = diag.loc[diag["Diagnosis"].isin(Gaming.good_diag)] |
| | |
| | ex["DatetimeHour"] = ex.apply(lambda x: Gaming.get_time_attributes(x, "Hour"), axis=1) |
| |
|
| | |
| | ex = ex.explode("DatetimeHour") |
| | ex["DatetimeHour"] = ex["DatetimeHour"].apply(lambda x: x.replace(minute=0, second=0, microsecond=0)) |
| | ex["HourCount"] = ex.groupby(["AppCode", "CreatedDatetime"])["DatetimeHour"].transform('count') |
| | |
| | ex["BlackoutHours"] = ex["BlackoutHours"] / ex["HourCount"] |
| |
|
| | |
| | ex = ex.groupby(["AppCode", "DatetimeHour"])["BlackoutHours"].sum().reset_index() |
| |
|
| | config_user_dict = serialize.open_yaml("config_user.yaml") |
| |
|
| | if config_user_dict['local']['test']==False: |
| | serialize.save_pickle(ex, os.path.join(Gaming.gaming_dir,file_name)) |
| | return ex |
| |
|
| | @staticmethod |
| | def get_time_attributes(df, kind): |
| | start = df["PrevCreatedDatetime"] |
| | end = df["CreatedDatetime"] |
| | if kind == "Hour": |
| | thing = [x for x in pd.date_range(start, end, freq="H")] |
| | else: |
| | thing = [x.weekday() for x in pd.date_range(start, end, freq="D")] |
| | return thing |
| |
|