Read plain or compressed log files from disk, deliver as [batches of] lines to a log consumer. Wait for the log consumer to report success. Advance bookmark. Repeat ad infinitum.
npm i safe-log-reader
const read = require('safe-line-reader');
read.createReader(filePath, {
batchLimit: 1024,
bookmark: {
dir: path.resolve('someDir', '.bookmark'),
}
})
.on('readable', function () { this.readLine(); })
.on('read', function (line, count) {
// do something with this line of text
})
.on('end', function (done) {
// close up shop and go home
});
- Read plain text files
- Handles common log file events
- reads growing files (aka: tail -F)
- reads rotated logs
- reads the new file when it appears
- fs.watch tested on:
- Mac OS X
- Linux
- FreeBSD
- fs.watch tested on:
- continues reading old log file until quiet
- reads the new file when it appears
- file truncation (echo '' > foo.log)
- watches for non-existent log to appear
- Read compressed log files
- gzip (zlib)
- bzip2
- Emits data as lines, upon request (paused mode streaming)
- Uses a Transform Stream to efficiently convert buffer streams to lines
- waits for confirmation, then advances bookmarks
- handles utf-8 multibyte characters properly
- Remembers previously read files (bookmarks)
- Perists across program restarts
- identifies files by inode
- saves file data: name, size, line count, inode
- When safe, uses byte position to efficiently resume reading
- Perists across program restarts
- cronolog naming syntax (/var/log/http/YYYY/MM/DD/access.log)
- watches existing directory ancestor
- winston naming syntax (app.log1, app.log2, etc.)
- zero dependencies
- log-ship-elastic-postfix
- reads batches of log entries
- parses with postfix-parser
- fetches matching docs from ES
- updates/creates normalized postfix docs
- saves docs to elasticsearch
- log-ship-elastic-qpsmtpd
- receives JSON parsed log lines
- saves to elasticsearch
- tail-stream has good options for reading a file and handling rotation, truncation, and resuming. It had no tests so I wrote them and most have been merged. Bugs remain (demonstrated with Travis-CI test integration) unresolved. The author offered a license in exchange for the tests but the GPL is problematic.
- tail-forever has character encoding detection and very basic file watching.
- always-tail
The key "missing" feature of the node "tail" libraries is the ability to resume correctly after the app has stopped reading (think: kill -9) in the middle of a file.
Because files are read as [chunks of] bytes and log entries are lines, resuming at the files last byte position is likely to be in the middle of a line, or even splitting a multi-byte character. Extra buffered bytes not yet emitted as lines are lost, unless at restart, one rewinds and replays the last full $bufferSize. Then the consuming app needs to have duplicate line detection and suppression.
The key to resuming reading a log file safely is tracking line numbers and the byte steam offset the consuming app has committed. When saving bookmarks, the file position advances to the byte offset coinciding with the byte position of the last line your application has safely commited.
Safe-log-reader uses a Transform Stream to convert the byte stream into
lines. This makes it dead simple to read compressed files by adding
a .pipe(ungzip())
into the stream.
Copyright 2015 by eFolder, Inc.