AnomaLog turns raw logs into reproducible, model-ready datasets for anomaly detection.
It is designed for research workflows where preprocessing is not incidental, but part of the experimental artifact. Parsing, template mining, labeling, and sequence construction are made explicit, composable, and reproducible.
Benchmark results in log anomaly detection often depend on hidden preprocessing decisions. AnomaLog surfaces those decisions and makes them first-class, enabling fair comparison and repeatable experiments.
Some typical use cases include:
- Comparing anomaly detectors under controlled preprocessing assumptions
- Running ablations over parsers or template miners
- Reproducing published benchmarks from raw logs
from anomalog.presets import bgl
from anomalog.representations import TemplatePhraseRepresentation
samples = (
bgl.build()
.group_by_entity()
.with_train_fraction(0.8)
.represent_with(TemplatePhraseRepresentation())
)This constructs model-ready features from raw logs with a fully specified, reproducible preprocessing pipeline.
pip install anomalogAnomaLog models preprocessing as a deterministic pipeline:
- Source - raw log ingestion and dataset sourcing
- Parsing - structured parsing into typed log records
- Templating - template mining and assignment
- Sequencing - grouping logs into windows
- Representation - model-ready feature extraction
The public API is centered on the fluent builder:
dataset = (
DatasetSpec(...)
.from_source(...)
.parse_with(...)
.label_with(...)
.template_with(...)
.build()
)That produces a templated dataset, which can then be grouped into sequences and converted into model-ready representations.
Our presets include popular log anomaly datasets like BGL and HDFS v1.
A typical workflow is:
- Load a preset dataset (e.g. BGL).
- Build structured logs with parsing and templating.
- Group logs into labeled sequences.
- Convert sequences into model-ready features.
>>> from anomalog import SplitLabel
>>> from anomalog.parsers import IdentityTemplateParser
>>> from anomalog.presets import bgl
>>> from anomalog.representations import TemplatePhraseRepresentation
>>> dataset = bgl.build()
# Presets are ordinary DatasetSpec objects, so preprocessing choices stay visible
>>> bgl.template_parser.name
'drain3'
# Deterministically group logs into sequences with explicit train/test semantics
>>> sequences = dataset.group_by_entity().with_train_fraction(0.8)
# Each event stores (template, parameters, timing delta)
>>> next(iter(sequences))
TemplateSequence(
events=[
('RAS KERNEL FATAL <:*:> <:*:> <:*:>', ['data', 'storage', 'interrupt'], None),
('RAS KERNEL FATAL <:*:> <:*:> <:*:>', ['instruction', 'address:', '0x00004ed8'], 523407),
...
],
label=1,
entity_ids=['R00-M0-N0-C:J08-U01'],
split_label=<SplitLabel.TRAIN: 'train'>
)
# Convert sequences into n-gram features for modeling
>>> train_samples = sequences.represent_with(
TemplatePhraseRepresentation(phrase_ngram_min=1, phrase_ngram_max=2),
)
>>> next(sample for sample in train_samples if sample.split_label is SplitLabel.TRAIN)
SequenceSample(data=Counter({'ras': 49, 'kernel': 49, 'ras kernel': 49, ...})
label=1,
entity_ids=['R00-M0-N0-C:J08-U01'],
split_label=<SplitLabel.TRAIN: 'train'>,
window_id=0)
# Ablation: disable template mining and use raw log lines directly
>>> ablated_dataset = bgl.template_with(IdentityTemplateParser).build()Built-in presets are ordinary DatasetSpec objects, so the exact source,
parser, label, and template-mining choices stay visible in code and can be
modified for ablations instead of being hidden behind preprocessed artifacts.
You can also define a custom dataset:
from pathlib import Path
from anomalog import DatasetSpec
from anomalog.labels import CSVReader
from anomalog.parsers import HDFSV1Parser
from anomalog.sources import LocalZipSource
dataset = (
DatasetSpec("my-hdfs")
.from_source(
LocalZipSource(
Path("HDFS_v1.zip"),
raw_logs_relpath=Path("HDFS.log"),
),
)
.parse_with(HDFSV1Parser())
.label_with(
CSVReader(
relative_path=Path("preprocessed/anomaly_label.csv"),
entity_column="BlockId",
label_column="Label",
),
)
.build()
)The full documentation is organised by task:
- Getting started for the end-to-end workflow, grouping choices, representation stage, and split semantics
- Experiments for config-driven detector runs and recorded artifacts
- Reference for the codebase map and API pages
The repository also includes a config-driven experiment layer under
experiments/ for model experimentation on top of AnomaLog preprocessing.
Built-in experiment detectors include Template Frequency, Naive Bayes, River, and a scoped DeepLog and DeepCASE reimplementation.
uv run python -m experiments.runners.run_experiment \
--config experiments/configs/runs/bgl_template_frequency.tomlExperiment runs reuse AnomaLog's dataset-side caches and write deterministic result directories, but detector training and test scoring are intentionally rerun for new config fingerprints.
See experiments/README.md for the experiment layout and artifact format.
Contributor setup and local commands are documented in Development.