Loghub-2.0 is a collection of large-scale annotated datasets for log parsing based on Loghub, proposed by LogPAI.
Based on Loghub-2.0, we propose a more comprehensive benchmark of log parsers. The detailed evaluation results could be found at RQ_experiments 🔗.
Software systems | # Annotated Logs (Loghub-2.0) | # Templates (Loghub-2.0) | # Templates (Loghub-2k) |
---|---|---|---|
Distributed systems | |||
Hadoop | 179,993 | 236 | 114 |
HDFS | 11,167,740 | 46 | 14 |
OpenStack | 207,632 | 48 | 43 |
Spark | 16,075,117 | 236 | 36 |
Zookeeper | 74,273 | 89 | 50 |
Supercomputer systems | |||
BGL | 4,631,261 | 320 | 120 |
HPC | 429,987 | 74 | 46 |
Thunderbird | 16,601,745 | 1,241 | 149 |
Operating systems | |||
Linux | 23,921 | 338 | 118 |
Mac | 100,314 | 626 | 341 |
Server application | |||
Apache | 51,977 | 29 | 6 |
OpenSSH | 638,946 | 38 | 27 |
Standalone software | |||
HealthApp | 212,394 | 156 | 75 |
Proxifier | 21,320 | 11 | 8 |
Average | 3,601,187 | 249.1 | 81.9 |
Please first download the full datasets of Loghub-2.0 via Zenodo.
Then, you need to put these datasets into full_dataset/
following the format of 2k_dataset
.
├── 2k_dataset/ # the original Loghub-2k datasets
├── full_dataset/ # unzip the Loghub-2.0 into this directory
│ └── post_process.py # we provide the heuristic roles used in our annotation of templates
├── benchmark/
│ ├── evaluation/
│ ├── logparser/
│ ├── old_benchmark/
│ ├── LogPPT/ # contains the modified source code of LogPPT
│ ├── UniParser/ # contains the source code of implemented UniParser
│ ├── run_statistic_2k.sh # the script to run all statistic-based log parsers on Loghub-2k datasets
│ └── run_statistic_full.sh # the script to run all statistic-based log parsers on Loghub-2.0 datasets
├── result/
│ ├── ...... #
│ └── ...... # contains the output evaluation metric files and all parsed results
├── RQ_experiments/ # contains the experimental results of RQs
│ ├── RQ1/
│ ├── RQ2/
│ └── RQ3/
├── requirements.txt
└── README.MD
Owing to the large scale of the benchmark in the experiments, the requirements of the benchmark of all log parsers are:
- At least 16GB memory.
- At least 100GB storage.
- GPU (for LogPPT and UniParser).
Installation
- Install
python >= 3.8
pip install -r requirements.txt
Running the entire benchmark using Loghub-2.0 datasets requires more than 48 hours to complete.
Note that if you would like to evaluate your parser, one can easily put their parsed results following the format as the files shown in result/
, and run our evluation code.
If you woud like to re-run all parsers using Loghub-2.0, please follow our large-scale benchmarking steps.
We give a demo script to run Drain on both Loghub-2k and Loghub-2.0, this will takes about 2-3 hours.
cd benchmark/
./demo.sh
One can follow the steps to evaluate all parsers using Loghub-2k or the proposed Loghub-2.0 datasets. The overall time cost is more than 48 hours.
- Run all statistic-based log parsers on Loghub-2k
cd benchmark/
./run_statistic_2k.sh
- Run all statistic-based log parsers on Loghub-2.0
cd benchmark/
./run_statistic_full.sh
-
Run Semantic-based log parsers: LogPPT & UniParser
Since these methods are quite different with other log parsers, and they requires a GPU to support efficient parsing, we seperate their environments from other log parsers. Please refer to the README file of LogPPT or UniParser to use one-click script to parse and evaluate each log parsers respectively.
If you use our benchmark or datasets for research, please cite the following papers:
-
Zhihan Jiang, Jinyang Liu, Junjie Huang, Yichen Li, Yintong Huo, Jiazhen Gu, Zhuangbin Chen, Jieming Zhu, Michael R. Lyu. A Large-scale Evaluation for Log Parsing Techniques: How Far are We? ISSTA, 2024.
-
Jieming Zhu, Shilin He, Pinjia He, Jinyang Liu, Michael R. Lyu. Loghub: A Large Collection of System Log Datasets for AI-driven Log Analytics. ISSRE, 2023.
In addition, if you use the souce code of our benchmark for research, please also cite the following two papers:
-
Khan Zanis Ali, Shin Donghwan, Bianculli Domenico, Briand Lionel. Guidelines for Assessing the Accuracy of Log Message Template Identification Techniques. ICSE, 2022.
-
Jieming Zhu, Shilin He, Jinyang Liu, Pinjia He, Qi Xie, Zibin Zheng, Michael R. Lyu. Tools and Benchmarks for Automated Log Parsing. ICSE, 2019.