Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Turing Change Point Detection Benchmark #167

Open
Tracked by #103
RobotPsychologist opened this issue Nov 19, 2024 · 2 comments
Open
Tracked by #103

Turing Change Point Detection Benchmark #167

RobotPsychologist opened this issue Nov 19, 2024 · 2 comments
Assignees
Labels
research Background research and research publication team.

Comments

@RobotPsychologist
Copy link
Owner

No description provided.

@sneha3799
Copy link
Contributor

Hello, I will take this one. (https://arxiv.org/pdf/2003.06222)

@sneha3799
Copy link
Contributor

The paper, An Evaluation of Change Point Detection Algorithms, focuses on comparing methods designed to identify "change points" in time series data—moments where patterns shift significantly, such as a sudden trend change or spike. The researchers introduced a novel dataset annotated by humans and designed a comprehensive evaluation framework incorporating two complementary metrics to benchmark algorithm performance.

Categories of CPD Methods
Online vs. Offline:
Online: Detects changes in real-time as data arrives.
Offline: Analyzes the entire dataset post-collection.
Univariate vs. Multivariate:
Univariate: Analyzes single-variable data.
Multivariate: Examines relationships in multi-variable datasets.
Model-Based vs. Nonparametric:
Model-Based: Relies on statistical models (e.g., Bayesian, frequentist).
Nonparametric: Uses principles like divergence measures, avoiding strict assumptions.

Performance Metrics
Evaluation involves two complementary perspectives:

Clustering Metrics: Measures segmentation quality using metrics like Variation of Information (VI) and the Segmentation Covering Metric (based on the Jaccard Index).
Classification Metrics: Assesses precision, recall, and F1-score for detecting individual change points.
Margin of Error: Allows a small tolerance (e.g., ±5 time units) to account for annotation discrepancies.

Annotation Process
Annotation Tool: A web-based application enabled annotators to mark change points, with training provided on synthetic data.
Dataset:
Comprises 42 time series (real-world and simulated).
Features include seasonality, abrupt changes, and outliers.
Quality Control:
Simulated datasets confirmed annotator accuracy.
High inter-annotator agreement (F1-score ~0.9) indicated strong consistency.

Experimental Setup
Methods Tested:
Included both traditional (e.g., CUSUM, binary segmentation) and advanced (e.g., Bayesian CPD) algorithms.
Evaluation Conditions:
Default: Algorithms tested with standard configurations.
Oracle: Hyperparameters optimized for best performance.
Metrics:
F1-score for point-level detection.
Segmentation covering metric for segment-level accuracy.

Key Findings
Algorithm Performance:
Default Settings: Binary segmentation performed well for univariate data.
Tuned Settings: Bayesian CPD (BOCPD) excelled, particularly for multivariate cases.
General Observations:
Simpler methods often matched or outperformed complex ones for low-change scenarios.
No single method dominated across all datasets and conditions.
Challenges:
Multivariate data posed greater difficulties.
Annotator disagreement occurred in cases of ambiguous patterns (e.g., periodicity).

Impact and Future Directions
Dataset Contribution:
Provides a robust, real-world benchmark for CPD research.
Inspired by datasets like Berkeley Segmentation Data Set in image analysis.
Open-Source Tools:
Annotation tool and experimental framework are publicly available.
Recommendations:
Develop automated hyperparameter tuning.
Investigate why simpler algorithms perform competitively without tuning.
This study establishes a foundation for evaluating CPD methods rigorously and encourages further advancements in the field.

The codes and data are available at https://github.com/alan-turing-institute/TCPDBench.
The annotation tool is a web application and is made available as open source software at https://github.com/alan-turing-institute/AnnotateChange

1
2

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
research Background research and research publication team.
Development

No branches or pull requests

2 participants