| import json |
| from pathlib import Path |
|
|
| import numpy as np |
| import fire |
| import h5py |
| from sklearn.metrics import accuracy_score, mean_absolute_error, f1_score |
|
|
| from infer_eval_utils import read_time_series_data, concat_base_path, non_zero_rel_mae, DATASET_TO_TASK |
|
|
|
|
| class Runner(object): |
|
|
| def multitask_classification(self, |
| infer_path: str = "", |
| gts: list = [], |
| preds: list = []): |
| tasks = gts[0].keys() |
| output_fpath = Path( |
| infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| res_dict = {} |
| success = 0 |
| fail = 0 |
| for task in tasks: |
| correct_count = 0 |
| for gt, pred in zip(gts, preds): |
| try: |
| if pred[task] == gt[task][0] or pred[task].lower( |
| ) == gt[task][0]: |
| correct_count += 1 |
| success += 1 |
| except: |
| fail += 1 |
| acc = correct_count / len(gts) |
| print(f"Accuracy for {task}: {acc}") |
| class_f1s = [] |
| task_gts = [gt[task][0] for gt in gts] |
| labels = list(set(task_gts)) |
| for label in labels: |
| |
| true_indices = [ |
| i for i, gt in enumerate(task_gts) if gt == label |
| ] |
| if len(true_indices) == 0: |
| recall = 0.0 |
| else: |
| correct_predictions = 0 |
| for idx in true_indices: |
| pred = preds[idx][task] |
| gt = task_gts[idx] |
| if pred == gt or pred.lower() == gt: |
| correct_predictions += 1 |
|
|
| recall = correct_predictions / len(true_indices) |
|
|
| pred_indices = [ |
| i for i, pred in enumerate(preds) |
| if pred[task].lower() == label or pred[task] == label |
| ] |
| if len(pred_indices) == 0: |
| precision = 0.0 |
| else: |
| correct_predictions = 0 |
| for idx in pred_indices: |
| pred = preds[idx][task] |
| gt = task_gts[idx] |
| if pred == gt or pred.lower() == gt: |
| correct_predictions += 1 |
| precision = correct_predictions / len(pred_indices) |
|
|
| f1 = (2 * recall * precision) / (recall + precision + 1e-6) |
| class_f1s.append(f1) |
|
|
| res_dict[task] = { |
| "acc": acc, |
| "f1": f1, |
| "success": success, |
| "fail": fail, |
| "success_rate": success / (success + fail) |
| } |
|
|
| res_dict["overall"] = { |
| "f1": np.mean(class_f1s), |
| "acc": np.mean([r["acc"] for r in res_dict.values()]), |
| } |
|
|
| with open(output_fpath, "w") as writer: |
| json.dump(res_dict, writer, indent=4) |
| writer.write("\n") |
|
|
| def multichoice_classification(self, |
| infer_path: str = "", |
| gts: list = [], |
| preds: list = []): |
| all_labels = set() |
| for gt in gts: |
| if isinstance(gt, list): |
| all_labels.update(gt) |
| else: |
| all_labels.add(gt) |
|
|
| all_labels = sorted(list(all_labels)) |
|
|
| |
| y_true_multilabel = [] |
| y_pred_multilabel = [] |
| success = 0 |
| fail = 0 |
| for gt, pred in zip(gts, preds): |
| |
| if isinstance(gt, list): |
| gt_labels = gt |
| else: |
| gt_labels = [gt] |
|
|
| |
| while '\n\n' in pred: |
| pred = pred.replace('\n\n', '\n') |
| pred_labels = [x.strip() for x in pred.split("\n")] |
|
|
| |
| gt_binary = [ |
| 1 if label in gt_labels else 0 for label in all_labels |
| ] |
| pred_binary = [] |
| for label in all_labels: |
| is_found = False |
| for pred_label in pred_labels: |
| if pred_label == label or pred_label.lower() == label: |
| is_found = True |
| break |
| if is_found: |
| pred_binary.append(1) |
| else: |
| pred_binary.append(0) |
|
|
| y_true_multilabel.append(gt_binary) |
| y_pred_multilabel.append(pred_binary) |
|
|
| y_true_multilabel = np.array(y_true_multilabel) |
| y_pred_multilabel = np.array(y_pred_multilabel) |
|
|
| |
| f1_scores = [] |
| for i, label in enumerate(all_labels): |
| f1 = f1_score(y_true_multilabel[:, i], |
| y_pred_multilabel[:, i], |
| zero_division=0) |
| f1_scores.append(f1) |
| print(f"F1 score for class {label}: {f1:.4f}") |
|
|
| |
| macro_f1 = np.mean(f1_scores) |
| print(f"Macro F1 score (mean of all classes): {macro_f1:.4f}") |
|
|
| |
| output_fpath = Path( |
| infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| output_fpath.parent.mkdir(parents=True, exist_ok=True) |
|
|
| results = { |
| "macro_f1": macro_f1, |
| "per_class_f1": dict(zip(all_labels, f1_scores)) |
| } |
|
|
| with open(output_fpath, "w") as writer: |
| json.dump(results, writer, indent=4) |
| writer.write("\n") |
|
|
| def classification(self, infer_path: str = ""): |
| gts, preds = [], [] |
| with open(infer_path, "r") as f: |
| for line in f: |
| item = json.loads(line) |
| if "id" not in item: |
| continue |
| gts.append(item["ground_truth"]) |
| preds.append(item["output"]) |
|
|
| if any(isinstance(gt, list) for gt in gts): |
| return self.multichoice_classification(infer_path, gts, preds) |
|
|
| if isinstance(preds[0], dict): |
| return self.multitask_classification(infer_path, gts, preds) |
| |
| |
| correct_count = 0 |
| for gt, pred in zip(gts, preds): |
| if pred == gt or pred.lower() == gt: |
| correct_count += 1 |
|
|
| acc = correct_count / len(gts) |
| print(f"Accuracy: {acc}") |
|
|
| |
| labels = list(set(gts)) |
|
|
| |
| class_recalls = [] |
| class_precisions = [] |
| class_f1s = [] |
| for label in labels: |
| |
| true_indices = [i for i, gt in enumerate(gts) if gt == label] |
| if len(true_indices) == 0: |
| recall = 0.0 |
| class_recalls.append(0.0) |
| else: |
| |
| correct_predictions = 0 |
| for idx in true_indices: |
| pred = preds[idx] |
| gt = gts[idx] |
| if pred == gt or pred.lower() == gt: |
| correct_predictions += 1 |
|
|
| recall = correct_predictions / len(true_indices) |
| class_recalls.append(recall) |
|
|
| pred_indices = [ |
| i for i, pred in enumerate(preds) |
| if pred.lower() == label or pred == label |
| ] |
| if len(pred_indices) == 0: |
| precision = 0.0 |
| class_precisions.append(0.0) |
| else: |
| correct_predictions = 0 |
| for idx in pred_indices: |
| pred = preds[idx] |
| gt = gts[idx] |
| if pred == gt or pred.lower() == gt: |
| correct_predictions += 1 |
|
|
| precision = correct_predictions / len(pred_indices) |
| class_precisions.append(precision) |
|
|
| f1 = (2 * recall * precision) / (recall + precision + 1e-6) |
| class_f1s.append(f1) |
|
|
| output_fpath = Path( |
| infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| with open(output_fpath, "w") as writer: |
| json.dump({ |
| "acc": acc, |
| "f1": np.mean(class_f1s), |
| }, |
| writer, |
| indent=4) |
| writer.write("\n") |
|
|
| def mcq(self, infer_path: str = ""): |
| gts, preds = [], [] |
| with open(infer_path, "r") as f: |
| for line in f: |
| item = json.loads(line) |
| if "id" not in item: |
| continue |
| gts.append(item["ground_truth"]) |
| preds.append(item["output"]) |
|
|
| |
| |
| correct_count = 0 |
| for gt, pred in zip(gts, preds): |
| if pred == gt or pred.lower() == gt: |
| correct_count += 1 |
|
|
| acc = correct_count / len(gts) |
| print(f"Accuracy: {acc}") |
|
|
| |
| |
| try: |
| labels = list(set(gts)) |
|
|
| |
| class_recalls = [] |
| for label in labels: |
| |
| true_indices = [i for i, gt in enumerate(gts) if gt == label] |
| if len(true_indices) == 0: |
| class_recalls.append(0.0) |
| continue |
|
|
| |
| correct_predictions = 0 |
| for idx in true_indices: |
| pred = preds[idx] |
| gt = gts[idx] |
| if pred == gt or pred.lower() == gt: |
| correct_predictions += 1 |
|
|
| recall = correct_predictions / len(true_indices) |
| class_recalls.append(recall) |
|
|
| |
| weighted_recall = np.mean(class_recalls) |
| print(f"Weighted Recall: {weighted_recall}") |
|
|
| output_fpath = Path( |
| infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| with open(output_fpath, "w") as writer: |
| json.dump({ |
| "acc": acc, |
| "uar": weighted_recall |
| }, |
| writer, |
| indent=4) |
| writer.write("\n") |
|
|
| except Exception as e: |
| print(f"Error calculating Weighted Recall: {e}") |
| print( |
| "Possible reasons: labels are not numeric or contain non-numeric labels" |
| ) |
|
|
| def anomaly_detection(self, infer_path: str = ""): |
| gts, preds = [], [] |
| success, fail = 0, 0 |
| with open(infer_path, "r") as f: |
| for line in f: |
| item = json.loads(line) |
| if "id" not in item: |
| continue |
|
|
| if item["output"].lower() == "yes": |
| preds.append(True) |
| elif item["output"].lower() == "no": |
| preds.append(False) |
| else: |
| fail += 1 |
| continue |
| success += 1 |
| gts.append(item["ground_truth"]) |
|
|
| correct_count = 0 |
| for gt, pred in zip(gts, preds): |
| if pred == gt: |
| correct_count += 1 |
| print(f"Success: {success}, Fail: {fail}") |
| acc = correct_count / len(gts) |
| print(f"Accuracy: {acc}") |
|
|
| f1 = f1_score(gts, preds) |
| print(f"F1 Score: {f1}") |
| output_fpath = Path( |
| infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| with open(output_fpath, "w") as writer: |
| json.dump({ |
| "acc": acc, |
| "f1": f1, |
| "success": success, |
| "fail": fail |
| }, |
| writer, |
| indent=4) |
| writer.write("\n") |
|
|
| def forecasting(self, infer_path: str = ""): |
| gt_arrs = [] |
| pred_arrs = [] |
| success = 0 |
| fail = 0 |
| with h5py.File(infer_path, "r") as f: |
| base_path = Path(f["base_path"][()].decode("utf-8")) |
| for id in f.keys(): |
| try: |
| if id not in [ |
| "base_path", "dataset_name", "domain", "task", |
| "scene" |
| ]: |
| gt_path = concat_base_path( |
| base_path, |
| f[id]["gt_path"][()].decode("utf-8").strip("/")) |
| gt_data = read_time_series_data(gt_path) |
| gt_data = np.array(gt_data, dtype=np.float32) |
| pred = f[id]["pred_result"][()] |
|
|
| if pred.shape != gt_data.shape: |
| raise ValueError( |
| f"Pred shape {pred.shape} does not match gt shape {gt_data.shape}" |
| ) |
| gt_arrs.append(gt_data.reshape(-1)) |
| pred_arrs.append(pred.reshape(-1)) |
| success += 1 |
| except Exception as e: |
| print(f"Error processing {id}: {e}") |
| fail += 1 |
|
|
| if len(gt_arrs) == 0: |
| mae = "N/A" |
| rel_mae = "N/A" |
| else: |
| gt_arrs = np.concatenate(gt_arrs) |
| pred_arrs = np.concatenate(pred_arrs) |
|
|
| |
| mae = mean_absolute_error(gt_arrs, pred_arrs) |
| rel_mae = non_zero_rel_mae(gt_arrs, pred_arrs) |
| print( |
| f"MAE: {mae}, REL_MAE: {rel_mae}, Success: {success}, Fail: {fail}" |
| ) |
| output_fpath = Path( |
| infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| with open(output_fpath, "w") as writer: |
| json.dump( |
| { |
| "rel_mae": rel_mae, |
| "mae": mae, |
| "success": success, |
| "fail": fail, |
| "success_rate": success / (success + fail) |
| }, |
| writer, |
| indent=4) |
| writer.write("\n") |
|
|
| def synthesize(self, infer_path: str = ""): |
| return self.forecasting(infer_path) |
|
|
| def imputation(self, infer_path: str = ""): |
| gt_arrs = [] |
| pred_arrs = [] |
| success = 0 |
| fail = 0 |
| with h5py.File(infer_path, "r") as f: |
| base_path = Path(f["base_path"][()].decode("utf-8")) |
| for id in f.keys(): |
| try: |
| if id not in [ |
| "base_path", "dataset_name", "domain", "task", |
| "scene" |
| ]: |
| |
| |
| gt_path = concat_base_path( |
| base_path, |
| f[id]["gt_path"][()].decode("utf-8").strip("/")) |
| gt_data = read_time_series_data(gt_path) |
|
|
| |
| |
| input_path = concat_base_path( |
| base_path, f[id]["input_ts_path"][( |
| )].decode("utf-8").strip("/")) |
| input_data = read_time_series_data(input_path) |
|
|
| pred_indices = np.where(input_data == "X")[0] |
| pred = f[id]["pred_result"][()] |
|
|
| pred = pred[pred_indices] |
| gt_data = gt_data[pred_indices] |
| if len(pred) != len(gt_data): |
| length_mismatch += 1 |
| else: |
| success += 1 |
| if len(pred) < len(gt_data): |
| pred = pred[:len(gt_data)] |
| if len(pred) > len(gt_data): |
| gt_data = gt_data[:len(pred)] |
| gt_arrs.append(gt_data) |
| pred_arrs.append(pred) |
| |
| except Exception as e: |
| print(f"Error processing {id}: {e}") |
| fail += 1 |
|
|
| gt_arrs = np.concatenate(gt_arrs) |
| pred_arrs = np.concatenate(pred_arrs) |
| |
| rel_mae = non_zero_rel_mae(gt_arrs, pred_arrs) |
| mae = mean_absolute_error(gt_arrs, pred_arrs) |
|
|
| print( |
| f"REL_MAE: {rel_mae}, MAE: {mae}, Success: {success}, Fail: {fail}" |
| ) |
| output_fpath = Path( |
| infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| with open(output_fpath, "w") as writer: |
| json.dump( |
| { |
| "rel_mae": rel_mae, |
| "mae": mae, |
| "success": success, |
| "fail": fail, |
| "success_rate": success / (success + fail) |
| }, |
| writer, |
| indent=4) |
| writer.write("\n") |
|
|
| def event_detection(self, infer_path: str = ""): |
| event_gts, event_preds = [], [] |
| seq_length = None |
| success = 0 |
| total = 0 |
| timestamp_gts, timestamp_preds = [], [] |
| with open(infer_path, "r") as f: |
| for line in f: |
| item = json.loads(line) |
| if "id" not in item: |
| seq_length = item["seq_length"] |
| continue |
| event_gt = item["ground_truth"]["contain"] |
| event_gts.append(1 if event_gt else 0) |
|
|
| if "\n" in item["output"]: |
| while '\n\n' in item["output"]: |
| item["output"] = item["output"].replace('\n\n', '\n') |
| event_pred, *timestamps = item["output"].split("\n") |
| else: |
| event_pred = item["output"] |
| timestamps = None |
| event_preds.append(1 if event_pred.lower() == "yes" else 0) |
|
|
| if event_gt: |
| if "start_time" in item["ground_truth"]: |
| gt_timestamps = [item["ground_truth"]["start_time"]] |
| elif "start_time_p" in item["ground_truth"]: |
| gt_timestamps = [ |
| item["ground_truth"]["start_time_p"], |
| item["ground_truth"]["start_time_s"] |
| ] |
|
|
| if timestamps is None: |
| pass |
| else: |
| try: |
| assert len(timestamps) == len(gt_timestamps) |
| for pred_timestamp, gt_timestamp in zip( |
| timestamps, gt_timestamps): |
| pred_timestamp = eval(pred_timestamp) |
| timestamp_preds.append(pred_timestamp) |
| timestamp_gts.append(gt_timestamp) |
| success += 1 |
| except Exception as e: |
| print(str(e)) |
| total += 1 |
|
|
| event_acc = accuracy_score(event_gts, event_preds) |
| event_f1 = f1_score(event_gts, event_preds) |
| timestamp_gts = np.array(timestamp_gts) |
| timestamp_preds = np.array(timestamp_preds) |
| mape = non_zero_rel_mae(timestamp_gts, timestamp_preds) |
| output_fpath = Path( |
| infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| with open(output_fpath, "w") as writer: |
| json.dump( |
| { |
| "acc": event_acc, |
| "f1": event_f1, |
| "mape": mape, |
| "success_rate": success / total |
| }, |
| writer, |
| indent=4) |
| writer.write("\n") |
| print({ |
| "acc": event_acc, |
| "f1": event_f1, |
| "mape": mape, |
| "success_rate": success / total |
| }) |
|
|
| def evaluate(self, infer_dir: str): |
| for infer_path in Path(infer_dir).glob("*"): |
| if infer_path.is_dir(): |
| continue |
| dataset_id = infer_path.stem |
| task = DATASET_TO_TASK[dataset_id] |
| print(f"evaluating {dataset_id} ...") |
| getattr(self, task)(infer_path) |
|
|
|
|
| if __name__ == "__main__": |
| fire.Fire(Runner) |
|
|