200 lines
12 KiB
Python
200 lines
12 KiB
Python
def get_script_language(script_path):
|
|
from pathlib import Path
|
|
script_path = Path(script_path)
|
|
if not script_path.exists():
|
|
raise ValueError("The following provider feature script does not exist: " + str(script_path))
|
|
|
|
if script_path.name.endswith(".py"):
|
|
return "python"
|
|
elif script_path.name.endswith(".R"):
|
|
return "r"
|
|
|
|
|
|
# Features.smk #########################################################################################################
|
|
def optional_phone_yield_input_for_locations(wildcards):
|
|
if config["PHONE_LOCATIONS"]["LOCATIONS_TO_USE"] in ["ALL_RESAMPLED","FUSED_RESAMPLED"]:
|
|
return "data/interim/{pid}/phone_yielded_timestamps.csv"
|
|
return []
|
|
|
|
def get_barnett_daily(wildcards):
|
|
if wildcards.provider_key.upper() == "BARNETT":
|
|
return "data/interim/{pid}/phone_locations_barnett_daily.csv"
|
|
return []
|
|
|
|
def get_locations_python_input(wildcards):
|
|
if wildcards.provider_key.upper() == "DORYAB":
|
|
return "data/interim/{pid}/phone_locations_processed_with_datetime_with_doryab_columns_episodes_resampled_with_datetime.csv"
|
|
else:
|
|
return "data/interim/{pid}/phone_locations_processed_with_datetime.csv"
|
|
|
|
def get_calls_input(wildcards):
|
|
if (wildcards.provider_key.upper() == "RAPIDS") and (config["PHONE_CALLS"]["PROVIDERS"]["RAPIDS"]["FEATURES_TYPE"] == "EPISODES"):
|
|
return "data/interim/{pid}/phone_calls_episodes_resampled_with_datetime.csv"
|
|
else:
|
|
return "data/raw/{pid}/phone_calls_with_datetime.csv"
|
|
|
|
def find_features_files(wildcards):
|
|
feature_files = []
|
|
for provider_key, provider in config[(wildcards.sensor_key).upper()]["PROVIDERS"].items():
|
|
if provider["COMPUTE"]:
|
|
feature_files.extend(expand("data/interim/{{pid}}/{sensor_key}_features/{sensor_key}_{language}_{provider_key}.csv", sensor_key=wildcards.sensor_key.lower(), language=get_script_language(provider["SRC_SCRIPT"]), provider_key=provider_key.lower()))
|
|
return(feature_files)
|
|
|
|
def find_empaticas_standardized_features_files(wildcards):
|
|
feature_files = []
|
|
if "empatica" in wildcards.sensor_key:
|
|
for provider_key, provider in config[(wildcards.sensor_key).upper()]["PROVIDERS"].items():
|
|
if provider["COMPUTE"] and provider.get("WINDOWS", False) and provider["WINDOWS"]["COMPUTE"]:
|
|
if "empatica" in wildcards.sensor_key:
|
|
feature_files.extend(expand("data/interim/{{pid}}/{sensor_key}_features/z_{sensor_key}_{language}_{provider_key}.csv", sensor_key=wildcards.sensor_key.lower(), language=get_script_language(provider["SRC_SCRIPT"]), provider_key=provider_key.lower()))
|
|
return(feature_files)
|
|
|
|
def find_joint_non_empatica_sensor_files(wildcards):
|
|
joined_files = []
|
|
for config_key in config.keys():
|
|
if config_key.startswith(("PHONE", "FITBIT")) and "PROVIDERS" in config[config_key] and isinstance(config[config_key]["PROVIDERS"], dict):
|
|
for provider_key, provider in config[config_key]["PROVIDERS"].items():
|
|
if "COMPUTE" in provider.keys() and provider["COMPUTE"]:
|
|
joined_files.append("data/processed/features/{pid}/" + config_key.lower() + ".csv")
|
|
break
|
|
return joined_files
|
|
|
|
|
|
def optional_steps_sleep_input(wildcards):
|
|
if config["FITBIT_STEPS_INTRADAY"]["EXCLUDE_SLEEP"]["FITBIT_BASED"]["EXCLUDE"]:
|
|
return "data/raw/{pid}/fitbit_sleep_summary_raw.csv"
|
|
else:
|
|
return []
|
|
|
|
def optional_steps_intraday_input(wildcards):
|
|
if config["FITBIT_STEPS_INTRADAY"]["EXCLUDE_SLEEP"]["TIME_BASED"]["EXCLUDE"] or config["FITBIT_STEPS_INTRADAY"]["EXCLUDE_SLEEP"]["FITBIT_BASED"]["EXCLUDE"]:
|
|
return "data/interim/{pid}/fitbit_steps_intraday_with_datetime_exclude_sleep.csv"
|
|
else:
|
|
return "data/raw/{pid}/fitbit_steps_intraday_with_datetime.csv"
|
|
|
|
def input_merge_sensor_features_for_individual_participants(wildcards):
|
|
feature_files = []
|
|
for config_key in config.keys():
|
|
if config_key.startswith(("PHONE", "FITBIT", "EMPATICA")) and "PROVIDERS" in config[config_key] and isinstance(config[config_key]["PROVIDERS"], dict):
|
|
for provider_key, provider in config[config_key]["PROVIDERS"].items():
|
|
if "COMPUTE" in provider.keys() and provider["COMPUTE"]:
|
|
feature_files.append("data/processed/features/{pid}/" + config_key.lower() + ".csv")
|
|
break
|
|
return feature_files
|
|
|
|
def input_merge_standardized_sensor_features_for_individual_participants(wildcards):
|
|
feature_files = []
|
|
for config_key in config.keys():
|
|
if config_key.startswith(("PHONE", "FITBIT", "EMPATICA")) and "PROVIDERS" in config[config_key] and isinstance(config[config_key]["PROVIDERS"], dict):
|
|
for provider_key, provider in config[config_key]["PROVIDERS"].items():
|
|
if "COMPUTE" in provider.keys() and provider["COMPUTE"] and ("STANDARDIZE_FEATURES" in provider.keys() and provider["STANDARDIZE_FEATURES"] or
|
|
"WINDOWS" in provider.keys() and "STANDARDIZE_FEATURES" in provider["WINDOWS"].keys() and provider["WINDOWS"]["STANDARDIZE_FEATURES"]):
|
|
feature_files.append("data/processed/features/{pid}/z_" + config_key.lower() + ".csv")
|
|
break
|
|
|
|
return feature_files
|
|
|
|
def get_phone_sensor_names():
|
|
phone_sensor_names = []
|
|
for config_key in config.keys():
|
|
if config_key.startswith(("PHONE")) and "PROVIDERS" in config[config_key]:
|
|
if config_key != "PHONE_DATA_YIELD" and config_key not in phone_sensor_names:
|
|
phone_sensor_names.append(config_key)
|
|
return phone_sensor_names
|
|
|
|
def pull_phone_data_input_with_mutation_scripts(wilcards):
|
|
from pathlib import Path
|
|
import yaml
|
|
input = dict()
|
|
phone_stream = config["PHONE_DATA_STREAMS"]["USE"]
|
|
|
|
input["participant_file"] = "data/external/participant_files/{pid}.yaml"
|
|
input["rapids_schema_file"] = "src/data/streams/rapids_columns.yaml"
|
|
input["stream_format"] = "src/data/streams/" + phone_stream + "/format.yaml"
|
|
|
|
if Path("src/data/streams/"+ phone_stream + "/container.R").exists():
|
|
input["stream_container"] = "src/data/streams/"+ phone_stream + "/container.R"
|
|
elif Path("src/data/streams/"+ phone_stream + "/container.py").exists():
|
|
input["stream_container"] = "src/data/streams/"+ phone_stream + "/container.py"
|
|
else:
|
|
raise ValueError("The container script for {stream} is missing: src/data/streams/{stream}/container.[py|R]".format(stream=empatica_stream))
|
|
|
|
schema = yaml.load(open(input.get("stream_format"), 'r'), Loader=yaml.FullLoader)
|
|
sensor = ("phone_" + wilcards.sensor).upper()
|
|
if sensor not in schema:
|
|
raise ValueError("{sensor} is not defined in the schema {schema}".format(sensor=sensor, schema=input.get("stream_format")))
|
|
|
|
for device_os in schema[sensor].keys():
|
|
if "MUTATION" not in schema[sensor][device_os]:
|
|
raise ValueError("MUTATION is missing from [{sensor}][{device_os}] of {schema}".format(sensor=sensor, device_os=device_os,schema=input.get("stream_format")))
|
|
if "COLUMN_MAPPINGS" not in schema[sensor][device_os]["MUTATION"]:
|
|
raise ValueError("COLUMN_MAPPINGS is missing from [{sensor}][{device_os}][MUTATION] of {schema}".format(sensor=sensor, device_os=device_os, schema=input.get("stream_format")))
|
|
if "SCRIPTS" not in schema[sensor][device_os]["MUTATION"]:
|
|
raise ValueError("SCRIPTS is missing from [{sensor}][{device_os}][MUTATION] of {schema}".format(sensor=sensor, device_os=device_os, schema=input.get("stream_format")))
|
|
|
|
scripts = schema[sensor][device_os]["MUTATION"]["SCRIPTS"]
|
|
if isinstance(scripts, list):
|
|
for idx, script in enumerate(scripts):
|
|
if not script.lower().endswith((".py", ".r")):
|
|
raise ValueError("Mutate scripts can only be Python or R scripts (.py, .R).\n Instead we got {script} in \n [{sensor}][{device_os}] of {schema}".format(script=script, sensor=sensor, device_os=device_os, schema=input.get("stream_format")))
|
|
input["mutationscript"+str(idx)] = script
|
|
return input
|
|
|
|
def input_tzcodes_file(wilcards):
|
|
from pathlib import Path
|
|
if config["TIMEZONE"]["TYPE"] == "MULTIPLE":
|
|
if not config["TIMEZONE"]["MULTIPLE"]["TZCODES_FILE"].lower().endswith(".csv"):
|
|
raise ValueError("[TIMEZONE][MULTIPLE][TZCODES_FILE] should point to a CSV file, instead you typed: " + config["TIMEZONE"]["MULTIPLE"]["TZCODES_FILE"])
|
|
if not Path(config["TIMEZONE"]["MULTIPLE"]["TZCODES_FILE"]).exists():
|
|
try:
|
|
config["TIMEZONE"]["MULTIPLE"]["TZ_FILE"]
|
|
except KeyError:
|
|
raise ValueError("To create TZCODES_FILE, a list of timezones should be created " +
|
|
"with the rule preprocessing.smk/prepare_tzcodes_file " +
|
|
"which will create a file specified as config['TIMEZONE']['MULTIPLE']['TZ_FILE']." +
|
|
"\n An alternative is to provide the file manually:" +
|
|
"[TIMEZONE][MULTIPLE][TZCODES_FILE] should point to a CSV file," +
|
|
"but the file in the path you typed does not exist: " +
|
|
config["TIMEZONE"]["MULTIPLE"]["TZCODES_FILE"])
|
|
return [config["TIMEZONE"]["MULTIPLE"]["TZCODES_FILE"]]
|
|
return []
|
|
|
|
def pull_wearable_data_input_with_mutation_scripts(wilcards):
|
|
import yaml
|
|
from pathlib import Path
|
|
input = dict()
|
|
device = wilcards.device_type.upper()
|
|
device_stream = config[device+"_DATA_STREAMS"]["USE"]
|
|
|
|
input["participant_file"] = "data/external/participant_files/{pid}.yaml"
|
|
input["rapids_schema_file"] = "src/data/streams/rapids_columns.yaml"
|
|
input["stream_format"] = "src/data/streams/" + device_stream + "/format.yaml"
|
|
|
|
if Path("src/data/streams/"+ device_stream + "/container.R").exists():
|
|
input["stream_container"] = "src/data/streams/"+ device_stream + "/container.R"
|
|
elif Path("src/data/streams/"+ device_stream + "/container.py").exists():
|
|
input["stream_container"] = "src/data/streams/"+ device_stream + "/container.py"
|
|
else:
|
|
raise ValueError("The container script for {stream} is missing: src/data/streams/{stream}/container.[py|R]".format(stream=device_stream))
|
|
|
|
schema = yaml.load(open(input.get("stream_format"), 'r'), Loader=yaml.FullLoader)
|
|
sensor = (device + "_" + wilcards.sensor).upper()
|
|
if sensor not in schema:
|
|
raise ValueError("{sensor} is not defined in the schema {schema}".format(sensor=sensor, schema=input.get("stream_format")))
|
|
|
|
if "MUTATION" not in schema[sensor]:
|
|
raise ValueError("MUTATION is missing from [{sensor}] of {schema}".format(sensor=sensor, schema=input.get("stream_format")))
|
|
if "COLUMN_MAPPINGS" not in schema[sensor]["MUTATION"]:
|
|
raise ValueError("COLUMN_MAPPINGS is missing from [{sensor}][MUTATION] of {schema}".format(sensor=sensor, schema=input.get("stream_format")))
|
|
if "SCRIPTS" not in schema[sensor]["MUTATION"]:
|
|
raise ValueError("SCRIPTS is missing from [{sensor}][MUTATION] of {schema}".format(sensor=sensor, schema=input.get("stream_format")))
|
|
|
|
scripts = schema[sensor]["MUTATION"]["SCRIPTS"]
|
|
if isinstance(scripts, list):
|
|
for idx, script in enumerate(scripts):
|
|
if not script.lower().endswith((".py", ".r")):
|
|
raise ValueError("Mutate scripts can only be Python or R scripts (.py, .R).\n Instead we got {script} in [{sensor}] of {schema}".format(script=script, sensor=sensor, schema=input.get("stream_format")))
|
|
input["mutationscript"+str(idx)] = script
|
|
return input
|
|
|