Add fitbit raw data and datetime
parent
ad514b5d40
commit
34c4586e4d
|
@ -7,6 +7,7 @@ include: "rules/reports.snakefile"
|
||||||
rule all:
|
rule all:
|
||||||
input:
|
input:
|
||||||
expand("data/raw/{pid}/{sensor}_raw.csv", pid=config["PIDS"], sensor=config["SENSORS"]),
|
expand("data/raw/{pid}/{sensor}_raw.csv", pid=config["PIDS"], sensor=config["SENSORS"]),
|
||||||
|
expand("data/raw/{pid}/{sensor}_raw.csv", pid=config["PIDS"], sensor=config["FITBIT_TABLE"]),
|
||||||
expand("data/raw/{pid}/{sensor}_with_datetime.csv", pid=config["PIDS"], sensor=config["SENSORS"]),
|
expand("data/raw/{pid}/{sensor}_with_datetime.csv", pid=config["PIDS"], sensor=config["SENSORS"]),
|
||||||
expand("data/processed/{pid}/battery_deltas.csv", pid=config["PIDS"]),
|
expand("data/processed/{pid}/battery_deltas.csv", pid=config["PIDS"]),
|
||||||
expand("data/processed/{pid}/screen_deltas.csv", pid=config["PIDS"]),
|
expand("data/processed/{pid}/screen_deltas.csv", pid=config["PIDS"]),
|
||||||
|
@ -39,6 +40,9 @@ rule all:
|
||||||
expand("data/processed/{pid}/accelerometer_{day_segment}.csv",
|
expand("data/processed/{pid}/accelerometer_{day_segment}.csv",
|
||||||
pid = config["PIDS"],
|
pid = config["PIDS"],
|
||||||
day_segment = config["ACCELEROMETER"]["DAY_SEGMENTS"]),
|
day_segment = config["ACCELEROMETER"]["DAY_SEGMENTS"]),
|
||||||
|
expand("data/raw/{pid}/fitbit_{fitbit_sensor}_with_datetime.csv",
|
||||||
|
pid=config["PIDS"],
|
||||||
|
fitbit_sensor=config["FITBIT_SENSORS"]),
|
||||||
# Reports
|
# Reports
|
||||||
expand("reports/figures/{pid}/{sensor}_heatmap_rows.html", pid=config["PIDS"], sensor=config["SENSORS"]),
|
expand("reports/figures/{pid}/{sensor}_heatmap_rows.html", pid=config["PIDS"], sensor=config["SENSORS"]),
|
||||||
expand("reports/figures/{pid}/compliance_heatmap.html", pid=config["PIDS"]),
|
expand("reports/figures/{pid}/compliance_heatmap.html", pid=config["PIDS"]),
|
||||||
|
|
|
@ -1,6 +1,9 @@
|
||||||
# Valid database table names
|
# Valid database table names
|
||||||
SENSORS: [applications_crashes, applications_foreground, applications_notifications, battery, bluetooth, calls, fitbit_data, locations, messages, plugin_ambient_noise, plugin_device_usage, plugin_google_activity_recognition, screen]
|
SENSORS: [applications_crashes, applications_foreground, applications_notifications, battery, bluetooth, calls, fitbit_data, locations, messages, plugin_ambient_noise, plugin_device_usage, plugin_google_activity_recognition, screen]
|
||||||
|
|
||||||
|
FITBIT_TABLE: [fitbit_data]
|
||||||
|
FITBIT_SENSORS: [heartrate, steps, sleep]
|
||||||
|
|
||||||
# Participants to include in the analysis
|
# Participants to include in the analysis
|
||||||
# You must create a file for each participant
|
# You must create a file for each participant
|
||||||
# named pXXX containing their device_id
|
# named pXXX containing their device_id
|
||||||
|
|
|
@ -15,6 +15,8 @@ rule readable_datetime:
|
||||||
params:
|
params:
|
||||||
timezones = None,
|
timezones = None,
|
||||||
fixed_timezone = config["READABLE_DATETIME"]["FIXED_TIMEZONE"]
|
fixed_timezone = config["READABLE_DATETIME"]["FIXED_TIMEZONE"]
|
||||||
|
wildcard_constraints:
|
||||||
|
sensor = "^fitbit.*" # ignoring fitbit sensors
|
||||||
output:
|
output:
|
||||||
"data/raw/{pid}/{sensor}_with_datetime.csv"
|
"data/raw/{pid}/{sensor}_with_datetime.csv"
|
||||||
script:
|
script:
|
||||||
|
@ -66,3 +68,33 @@ rule resample_fused_location:
|
||||||
"data/raw/{pid}/locations_resampled.csv"
|
"data/raw/{pid}/locations_resampled.csv"
|
||||||
script:
|
script:
|
||||||
"../src/data/resample_fused_location.R"
|
"../src/data/resample_fused_location.R"
|
||||||
|
|
||||||
|
rule fitbit_heartrate_with_datetime:
|
||||||
|
input:
|
||||||
|
"data/raw/{pid}/fitbit_data_raw.csv"
|
||||||
|
params:
|
||||||
|
local_timezone = config["READABLE_DATETIME"]["FIXED_TIMEZONE"],
|
||||||
|
output:
|
||||||
|
"data/raw/{pid}/fitbit_heartrate_with_datetime.csv"
|
||||||
|
script:
|
||||||
|
"../src/data/fitbit_heartrate_with_datetime.py"
|
||||||
|
|
||||||
|
rule fitbit_steps_with_datetime:
|
||||||
|
input:
|
||||||
|
"data/raw/{pid}/fitbit_data_raw.csv"
|
||||||
|
params:
|
||||||
|
local_timezone = config["READABLE_DATETIME"]["FIXED_TIMEZONE"]
|
||||||
|
output:
|
||||||
|
"data/raw/{pid}/fitbit_steps_with_datetime.csv"
|
||||||
|
script:
|
||||||
|
"../src/data/fitbit_steps_with_datetime.py"
|
||||||
|
|
||||||
|
rule fitbit_sleep_with_datetime:
|
||||||
|
input:
|
||||||
|
"data/raw/{pid}/fitbit_data_raw.csv"
|
||||||
|
params:
|
||||||
|
local_timezone = config["READABLE_DATETIME"]["FIXED_TIMEZONE"]
|
||||||
|
output:
|
||||||
|
"data/raw/{pid}/fitbit_sleep_with_datetime.csv"
|
||||||
|
script:
|
||||||
|
"../src/data/fitbit_sleep_with_datetime.py"
|
|
@ -0,0 +1,71 @@
|
||||||
|
import pandas as pd
|
||||||
|
import pytz, json
|
||||||
|
from datetime import datetime
|
||||||
|
|
||||||
|
|
||||||
|
NIGHT = "night"
|
||||||
|
MORNING = "morning"
|
||||||
|
AFTERNOON = "afternoon"
|
||||||
|
EVENING = "evening"
|
||||||
|
HOUR2EPOCH = [NIGHT] * 6 + [MORNING] * 6 + [AFTERNOON] * 6 + [EVENING] * 6
|
||||||
|
|
||||||
|
|
||||||
|
HR_COLUMNS = ("device_id",
|
||||||
|
"heartrate",
|
||||||
|
"local_date_time",
|
||||||
|
"local_date",
|
||||||
|
"local_month",
|
||||||
|
"local_day",
|
||||||
|
"local_day_of_week",
|
||||||
|
"local_time",
|
||||||
|
"local_hour",
|
||||||
|
"local_minute",
|
||||||
|
"local_day_segment")
|
||||||
|
|
||||||
|
fitbit_data = pd.read_csv(snakemake.input[0])
|
||||||
|
heartrate_data = fitbit_data[fitbit_data["fitbit_data_type"] == "heartrate"]
|
||||||
|
|
||||||
|
local_timezone = pytz.timezone(snakemake.params["local_timezone"])
|
||||||
|
|
||||||
|
|
||||||
|
"""
|
||||||
|
Data is pulled in intraday manner. Since data will be duplicated until the
|
||||||
|
last record from that day, first sort by time, then drop all but
|
||||||
|
the last record for each day. Drop duplicates based on aware timestamp.
|
||||||
|
"""
|
||||||
|
local_date_col = heartrate_data["timestamp"].apply(lambda ts: str(datetime.fromtimestamp(ts/1000, tz=local_timezone).date()))
|
||||||
|
heartrate_data = heartrate_data.assign(local_date=local_date_col.values)
|
||||||
|
heartrate_data.sort_values(by="timestamp", ascending=True, inplace=True)
|
||||||
|
heartrate_data.drop_duplicates(subset="local_date", keep="last", inplace=True)
|
||||||
|
|
||||||
|
device_id = heartrate_data["device_id"].iloc[0]
|
||||||
|
records = []
|
||||||
|
# Parse JSON into individual records
|
||||||
|
for record in heartrate_data.fitbit_data:
|
||||||
|
record = json.loads(record) # Parse text into JSON
|
||||||
|
curr_date = datetime.strptime(record["activities-heart"][0]["dateTime"], "%Y-%m-%d")
|
||||||
|
dataset = record["activities-heart-intraday"]["dataset"]
|
||||||
|
for data in dataset:
|
||||||
|
d_time = datetime.strptime(data["time"], '%H:%M:%S').time()
|
||||||
|
d_datetime = datetime.combine(curr_date, d_time)
|
||||||
|
|
||||||
|
# Create tuple of parsed data
|
||||||
|
row = (device_id,
|
||||||
|
data["value"],
|
||||||
|
d_datetime,
|
||||||
|
d_datetime.date(),
|
||||||
|
d_datetime.month,
|
||||||
|
d_datetime.day,
|
||||||
|
d_datetime.weekday(),
|
||||||
|
d_datetime.time(),
|
||||||
|
d_datetime.hour,
|
||||||
|
d_datetime.minute,
|
||||||
|
HOUR2EPOCH[d_datetime.hour])
|
||||||
|
|
||||||
|
# Append the data to a list
|
||||||
|
records.append(row)
|
||||||
|
|
||||||
|
# Create a new DataFrame from the list of tuples.
|
||||||
|
heartrate_preprocessed = pd.DataFrame(data=records, columns=HR_COLUMNS)
|
||||||
|
|
||||||
|
heartrate_preprocessed.to_csv(snakemake.output[0], index=False)
|
|
@ -0,0 +1,76 @@
|
||||||
|
import pandas as pd
|
||||||
|
import pytz, json
|
||||||
|
from datetime import datetime
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
NIGHT = "night"
|
||||||
|
MORNING = "morning"
|
||||||
|
AFTERNOON = "afternoon"
|
||||||
|
EVENING = "evening"
|
||||||
|
HOUR2EPOCH = [NIGHT] * 6 + [MORNING] * 6 + [AFTERNOON] * 6 + [EVENING] * 6
|
||||||
|
|
||||||
|
|
||||||
|
SLEEP_COLUMNS = ("device_id",
|
||||||
|
"sleep", # 1: "asleep", 2: "restless", or 3: "awake"
|
||||||
|
"local_date_time",
|
||||||
|
"local_date",
|
||||||
|
"local_month",
|
||||||
|
"local_day",
|
||||||
|
"local_day_of_week",
|
||||||
|
"local_time",
|
||||||
|
"local_hour",
|
||||||
|
"local_minute",
|
||||||
|
"local_day_segment")
|
||||||
|
|
||||||
|
fitbit_data = pd.read_csv(snakemake.input[0])
|
||||||
|
sleep_data = fitbit_data[fitbit_data["fitbit_data_type"] == "sleep"]
|
||||||
|
|
||||||
|
local_timezone = pytz.timezone(snakemake.params["local_timezone"])
|
||||||
|
|
||||||
|
|
||||||
|
"""
|
||||||
|
Data is pulled in intraday manner. Since data will be duplicated until the
|
||||||
|
last record from that day, first sort by time, then drop all but
|
||||||
|
the last record for each day. Drop duplicates based on aware timestamp.
|
||||||
|
"""
|
||||||
|
local_date_col = sleep_data["timestamp"].apply(lambda ts: str(datetime.fromtimestamp(ts/1000, tz=local_timezone).date()))
|
||||||
|
sleep_data = sleep_data.assign(local_date=local_date_col.values)
|
||||||
|
sleep_data.sort_values(by="timestamp", ascending=True, inplace=True)
|
||||||
|
sleep_data.drop_duplicates(subset="local_date", keep="last", inplace=True)
|
||||||
|
|
||||||
|
device_id = sleep_data["device_id"].iloc[0]
|
||||||
|
records = []
|
||||||
|
# Parse JSON into individual records
|
||||||
|
for multi_record in sleep_data.fitbit_data:
|
||||||
|
for record in json.loads(multi_record)["sleep"]:
|
||||||
|
start_date = datetime.strptime(record["startTime"][:10], "%Y-%m-%d")
|
||||||
|
end_date = datetime.strptime(record["endTime"][:10], "%Y-%m-%d")
|
||||||
|
flag = 1 if start_date == end_date else 0
|
||||||
|
for data in record["minuteData"]:
|
||||||
|
d_time = datetime.strptime(data["dateTime"], '%H:%M:%S').time()
|
||||||
|
if not flag and not d_time.hour:
|
||||||
|
flag = 1
|
||||||
|
curr_date = end_date if flag else start_date
|
||||||
|
d_datetime = datetime.combine(curr_date, d_time)
|
||||||
|
|
||||||
|
# Create tuple of parsed data
|
||||||
|
row = (device_id,
|
||||||
|
data["value"],
|
||||||
|
d_datetime,
|
||||||
|
d_datetime.date(),
|
||||||
|
d_datetime.month,
|
||||||
|
d_datetime.day,
|
||||||
|
d_datetime.weekday(),
|
||||||
|
d_datetime.time(),
|
||||||
|
d_datetime.hour,
|
||||||
|
d_datetime.minute,
|
||||||
|
HOUR2EPOCH[d_datetime.hour])
|
||||||
|
|
||||||
|
# Append the data to a list
|
||||||
|
records.append(row)
|
||||||
|
|
||||||
|
# Create a new DataFrame from the list of tuples.
|
||||||
|
sleep_preprocessed = pd.DataFrame(data=records, columns=SLEEP_COLUMNS)
|
||||||
|
|
||||||
|
sleep_preprocessed.to_csv(snakemake.output[0], index=False)
|
|
@ -0,0 +1,72 @@
|
||||||
|
import pandas as pd
|
||||||
|
import pytz, json
|
||||||
|
from datetime import datetime
|
||||||
|
|
||||||
|
|
||||||
|
NIGHT = "night"
|
||||||
|
MORNING = "morning"
|
||||||
|
AFTERNOON = "afternoon"
|
||||||
|
EVENING = "evening"
|
||||||
|
HOUR2EPOCH = [NIGHT] * 6 + [MORNING] * 6 + [AFTERNOON] * 6 + [EVENING] * 6
|
||||||
|
|
||||||
|
|
||||||
|
STEPS_COLUMNS = ("device_id",
|
||||||
|
"steps",
|
||||||
|
"local_date_time",
|
||||||
|
"local_date",
|
||||||
|
"local_month",
|
||||||
|
"local_day",
|
||||||
|
"local_day_of_week",
|
||||||
|
"local_time",
|
||||||
|
"local_hour",
|
||||||
|
"local_minute",
|
||||||
|
"local_day_segment")
|
||||||
|
|
||||||
|
fitbit_data = pd.read_csv(snakemake.input[0])
|
||||||
|
steps_data = fitbit_data[fitbit_data["fitbit_data_type"] == "steps"]
|
||||||
|
|
||||||
|
local_timezone = pytz.timezone(snakemake.params["local_timezone"])
|
||||||
|
|
||||||
|
|
||||||
|
"""
|
||||||
|
Data is pulled in intraday manner. Since data will be duplicated until the
|
||||||
|
last record from that day, first sort by time, then drop all but
|
||||||
|
the last record for each day. Drop duplicates based on aware timestamp.
|
||||||
|
"""
|
||||||
|
local_date_col = steps_data["timestamp"].apply(lambda ts: str(datetime.fromtimestamp(ts/1000, tz=local_timezone).date()))
|
||||||
|
steps_data = steps_data.assign(local_date=local_date_col.values)
|
||||||
|
steps_data.sort_values(by="timestamp", ascending=True, inplace=True)
|
||||||
|
steps_data.drop_duplicates(subset="local_date", keep="last", inplace=True)
|
||||||
|
|
||||||
|
device_id = steps_data["device_id"].iloc[0]
|
||||||
|
records = []
|
||||||
|
# Parse JSON into individual records
|
||||||
|
for record in steps_data.fitbit_data:
|
||||||
|
record = json.loads(record) # Parse text into JSON
|
||||||
|
curr_date = datetime.strptime(
|
||||||
|
record["activities-steps"][0]["dateTime"], "%Y-%m-%d")
|
||||||
|
dataset = record["activities-steps-intraday"]["dataset"]
|
||||||
|
for data in dataset:
|
||||||
|
d_time = datetime.strptime(data["time"], '%H:%M:%S').time()
|
||||||
|
d_datetime = datetime.combine(curr_date, d_time)
|
||||||
|
|
||||||
|
# Create tuple of parsed data
|
||||||
|
row = (device_id,
|
||||||
|
data["value"],
|
||||||
|
d_datetime,
|
||||||
|
d_datetime.date(),
|
||||||
|
d_datetime.month,
|
||||||
|
d_datetime.day,
|
||||||
|
d_datetime.weekday(),
|
||||||
|
d_datetime.time(),
|
||||||
|
d_datetime.hour,
|
||||||
|
d_datetime.minute,
|
||||||
|
HOUR2EPOCH[d_datetime.hour])
|
||||||
|
|
||||||
|
# Append the data to a list
|
||||||
|
records.append(row)
|
||||||
|
|
||||||
|
# Create a new DataFrame from the list of tuples.
|
||||||
|
steps_preprocessed = pd.DataFrame(data=records, columns=STEPS_COLUMNS)
|
||||||
|
|
||||||
|
steps_preprocessed.to_csv(snakemake.output[0], index=False)
|
Loading…
Reference in New Issue