SIEM & Analytics Integration¶
Export blazehash output directly into Elastic, Splunk, threat intel platforms, or your data warehouse.
ECS NDJSON for Elastic / Splunk¶
Produces one Elastic Common Schema record per file as newline-delimited JSON. Compatible with Filebeat, Logstash, and Splunk HEC.
Filebeat config to ingest:
filebeat.inputs:
- type: log
enabled: true
paths:
- /path/to/evidence.ndjson
json.keys_under_root: true
json.add_error_key: true
output.elasticsearch:
hosts: ["https://your-es-cluster:9200"]
index: "blazehash-%{+yyyy.MM.dd}"
Splunk HEC ingestion:
curl -k https://splunk:8088/services/collector/raw \
-H "Authorization: Splunk YOUR_HEC_TOKEN" \
-d @evidence.ndjson
STIX 2.1 for Threat Intel Platforms¶
Produces a STIX 2.1 JSON Bundle with file and observed-data objects. Ready for ingestion into:
- MISP -- import as STIX 2.1 bundle
- OpenCTI -- import via STIX connector
- ThreatConnect, Recorded Future, and any OASIS STIX-compatible platform
Each file becomes a STIX file object with hash indicators, making it immediately searchable across your threat intel ecosystem.
Parquet for Data Lakes and Analytics¶
Apache Parquet columnar format. Efficient for large-scale analytics.
Query with DuckDB:
SELECT filename, sha256, size
FROM read_parquet('evidence.parquet')
WHERE size > 1000000
ORDER BY size DESC;
Query with pandas:
import pandas as pd
df = pd.read_parquet('evidence.parquet')
df[df['size'] > 1_000_000].sort_values('size', ascending=False)
Query with polars:
import polars as pl
pl.scan_parquet('evidence.parquet').filter(pl.col('size') > 1_000_000).collect()
DuckDB Output¶
Writes directly to a DuckDB database file. Query immediately:
SQLite Output¶
Query with any SQLite client:
Join against your own tables, build custom reports, or use as input for other tools.
JSON and JSONL¶
# JSON array (entire result set)
blazehash -r /mnt/evidence --format json -o evidence.json
# JSONL (one JSON object per line, streamable)
blazehash -r /mnt/evidence --format jsonl -o evidence.jsonl
JSONL is preferred for streaming ingestion (each line is independently parseable). JSON is convenient for smaller result sets or tools that expect an array.
CSV¶
Opens in Excel, Google Sheets, or any spreadsheet tool.
DFXML¶
Digital Forensics XML. Compatible with Autopsy, The Sleuth Kit, and other forensic platforms that support the Garfinkel DFXML standard.
sha256sum / md5sum compatible¶
blazehash -r /mnt/evidence -c sha256 --format sha256sum -o hashes.sha256
blazehash -r /mnt/evidence -c md5 --format md5sum -o hashes.md5
Verifiable with standard coreutils:
Combine with chain of custody¶
All output formats work with case metadata:
blazehash -r /mnt/evidence -c sha256 \
--case "CASE-2026-001" --examiner "Jane Smith" \
--format ecs -o evidence.ndjson
Case ID and examiner name are embedded in each record, making them searchable in your SIEM.