1
0
Fork 0
mirror of https://github.com/moby/moby.git synced 2022-11-09 12:21:53 -05:00

Add func to get an io.Reader for tail operations

Signed-off-by: Brian Goff <cpuguy83@gmail.com>
This commit is contained in:
Brian Goff 2018-04-05 12:38:06 -04:00
parent e78ec2aab6
commit 874867d4e5
2 changed files with 384 additions and 49 deletions

View file

@ -3,7 +3,9 @@
package tailfile // import "github.com/docker/docker/pkg/tailfile" package tailfile // import "github.com/docker/docker/pkg/tailfile"
import ( import (
"bufio"
"bytes" "bytes"
"context"
"errors" "errors"
"io" "io"
"os" "os"
@ -16,51 +18,205 @@ var eol = []byte("\n")
// ErrNonPositiveLinesNumber is an error returned if the lines number was negative. // ErrNonPositiveLinesNumber is an error returned if the lines number was negative.
var ErrNonPositiveLinesNumber = errors.New("The number of lines to extract from the file must be positive") var ErrNonPositiveLinesNumber = errors.New("The number of lines to extract from the file must be positive")
//TailFile returns last n lines of reader f (could be a nil). //TailFile returns last n lines of the passed in file.
func TailFile(f io.ReadSeeker, n int) ([][]byte, error) { func TailFile(f *os.File, n int) ([][]byte, error) {
if n <= 0 { size, err := f.Seek(0, io.SeekEnd)
return nil, ErrNonPositiveLinesNumber
}
size, err := f.Seek(0, os.SEEK_END)
if err != nil { if err != nil {
return nil, err return nil, err
} }
block := -1
var data []byte rAt := io.NewSectionReader(f, 0, size)
var cnt int r, nLines, err := NewTailReader(context.Background(), rAt, n)
for { if err != nil {
var b []byte
step := int64(block * blockSize)
left := size + step // how many bytes to beginning
if left < 0 {
if _, err := f.Seek(0, os.SEEK_SET); err != nil {
return nil, err return nil, err
} }
b = make([]byte, blockSize+left)
if _, err := f.Read(b); err != nil { buf := make([][]byte, 0, nLines)
return nil, err scanner := bufio.NewScanner(r)
for scanner.Scan() {
buf = append(buf, scanner.Bytes())
}
return buf, nil
}
// SizeReaderAt is an interface used to get a ReaderAt as well as the size of the underlying reader.
// Note that the size of the underlying reader should not change when using this interface.
type SizeReaderAt interface {
io.ReaderAt
Size() int64
}
// NewTailReader scopes the passed in reader to just the last N lines passed in
func NewTailReader(ctx context.Context, r SizeReaderAt, reqLines int) (io.Reader, int, error) {
return NewTailReaderWithDelimiter(ctx, r, reqLines, eol)
}
// NewTailReaderWithDelimiter scopes the passed in reader to just the last N lines passed in
// In this case a "line" is defined by the passed in delimiter.
//
// Delimiter lengths should be generally small, no more than 12 bytes
func NewTailReaderWithDelimiter(ctx context.Context, r SizeReaderAt, reqLines int, delimiter []byte) (io.Reader, int, error) {
if reqLines < 1 {
return nil, 0, ErrNonPositiveLinesNumber
}
if len(delimiter) == 0 {
return nil, 0, errors.New("must provide a delimiter")
}
var (
size = r.Size()
tailStart int64
tailEnd = size
found int
)
if int64(len(delimiter)) >= size {
return bytes.NewReader(nil), 0, nil
}
scanner := newScanner(r, delimiter)
for scanner.Scan(ctx) {
if err := scanner.Err(); err != nil {
return nil, 0, scanner.Err()
}
found++
if found == 1 {
tailEnd = scanner.End()
}
if found == reqLines {
break
}
}
tailStart = scanner.Start(ctx)
if found == 0 {
return bytes.NewReader(nil), 0, nil
}
if found < reqLines && tailStart != 0 {
tailStart = 0
}
return io.NewSectionReader(r, tailStart, tailEnd-tailStart), found, nil
}
func newScanner(r SizeReaderAt, delim []byte) *scanner {
size := r.Size()
readSize := blockSize
if readSize > int(size) {
readSize = int(size)
}
// silly case...
if len(delim) >= readSize/2 {
readSize = len(delim)*2 + 2
}
return &scanner{
r: r,
pos: size,
buf: make([]byte, readSize),
delim: delim,
}
}
type scanner struct {
r SizeReaderAt
pos int64
buf []byte
delim []byte
err error
idx int
done bool
}
func (s *scanner) Start(ctx context.Context) int64 {
if s.idx > 0 {
idx := bytes.LastIndex(s.buf[:s.idx], s.delim)
if idx >= 0 {
return s.pos + int64(idx) + int64(len(s.delim))
}
}
// slow path
buf := make([]byte, len(s.buf))
copy(buf, s.buf)
readAhead := &scanner{
r: s.r,
pos: s.pos,
delim: s.delim,
idx: s.idx,
buf: buf,
}
if !readAhead.Scan(ctx) {
return 0
}
return readAhead.End()
}
func (s *scanner) End() int64 {
return s.pos + int64(s.idx) + int64(len(s.delim))
}
func (s *scanner) Err() error {
return s.err
}
func (s *scanner) Scan(ctx context.Context) bool {
if s.err != nil {
return false
}
for {
select {
case <-ctx.Done():
s.err = ctx.Err()
return false
default:
}
idx := s.idx - len(s.delim)
if idx < 0 {
readSize := int(s.pos)
if readSize > len(s.buf) {
readSize = len(s.buf)
}
if readSize < len(s.delim) {
return false
}
offset := s.pos - int64(readSize)
n, err := s.r.ReadAt(s.buf[:readSize], offset)
if err != nil && err != io.EOF {
s.err = err
return false
}
s.pos -= int64(n)
idx = n
}
s.idx = bytes.LastIndex(s.buf[:idx], s.delim)
if s.idx >= 0 {
return true
}
if len(s.delim) > 1 && s.pos > 0 {
// in this case, there may be a partial delimiter at the front of the buffer, so set the position forward
// up to the maximum size partial that could be there so it can be read again in the next iteration with any
// potential remainder.
// An example where delimiter is `####`:
// [##asdfqwerty]
// ^
// This resets the position to where the arrow is pointing.
// It could actually check if a partial exists and at the front, but that is pretty similar to the indexing
// code above though a bit more complex since each byte has to be checked (`len(delimiter)-1`) factorial).
// It's much simpler and cleaner to just re-read `len(delimiter)-1` bytes again.
s.pos += int64(len(s.delim)) - 1
}
} }
data = append(b, data...)
break
} else {
b = make([]byte, blockSize)
if _, err := f.Seek(left, os.SEEK_SET); err != nil {
return nil, err
}
if _, err := f.Read(b); err != nil {
return nil, err
}
data = append(b, data...)
}
cnt += bytes.Count(b, eol)
if cnt > n {
break
}
block--
}
lines := bytes.Split(data, eol)
if n < len(lines) {
return lines[len(lines)-n-1 : len(lines)-1], nil
}
return lines[:len(lines)-1], nil
} }

View file

@ -1,9 +1,17 @@
package tailfile // import "github.com/docker/docker/pkg/tailfile" package tailfile // import "github.com/docker/docker/pkg/tailfile"
import ( import (
"bufio"
"bytes"
"context"
"fmt"
"io"
"io/ioutil" "io/ioutil"
"os" "os"
"strings"
"testing" "testing"
"gotest.tools/assert"
) )
func TestTailFile(t *testing.T) { func TestTailFile(t *testing.T) {
@ -42,7 +50,7 @@ truncated line`)
if _, err := f.Write(testFile); err != nil { if _, err := f.Write(testFile); err != nil {
t.Fatal(err) t.Fatal(err)
} }
if _, err := f.Seek(0, os.SEEK_SET); err != nil { if _, err := f.Seek(0, io.SeekStart); err != nil {
t.Fatal(err) t.Fatal(err)
} }
expected := []string{"last fourth line", "last fifth line"} expected := []string{"last fourth line", "last fifth line"}
@ -50,10 +58,12 @@ truncated line`)
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
if len(res) != len(expected) {
t.Fatalf("\nexpected:\n%s\n\nactual:\n%s", expected, res)
}
for i, l := range res { for i, l := range res {
t.Logf("%s", l)
if expected[i] != string(l) { if expected[i] != string(l) {
t.Fatalf("Expected line %s, got %s", expected[i], l) t.Fatalf("Expected line %q, got %q", expected[i], l)
} }
} }
} }
@ -71,7 +81,7 @@ truncated line`)
if _, err := f.Write(testFile); err != nil { if _, err := f.Write(testFile); err != nil {
t.Fatal(err) t.Fatal(err)
} }
if _, err := f.Seek(0, os.SEEK_SET); err != nil { if _, err := f.Seek(0, io.SeekStart); err != nil {
t.Fatal(err) t.Fatal(err)
} }
expected := []string{"first line", "second line"} expected := []string{"first line", "second line"}
@ -79,8 +89,10 @@ truncated line`)
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
if len(expected) != len(res) {
t.Fatalf("\nexpected:\n%s\n\nactual:\n%s", expected, res)
}
for i, l := range res { for i, l := range res {
t.Logf("%s", l)
if expected[i] != string(l) { if expected[i] != string(l) {
t.Fatalf("Expected line %s, got %s", expected[i], l) t.Fatalf("Expected line %s, got %s", expected[i], l)
} }
@ -116,11 +128,11 @@ truncated line`)
if _, err := f.Write(testFile); err != nil { if _, err := f.Write(testFile); err != nil {
t.Fatal(err) t.Fatal(err)
} }
if _, err := f.Seek(0, os.SEEK_SET); err != nil { if _, err := f.Seek(0, io.SeekStart); err != nil {
t.Fatal(err) t.Fatal(err)
} }
if _, err := TailFile(f, -1); err != ErrNonPositiveLinesNumber { if _, err := TailFile(f, -1); err != ErrNonPositiveLinesNumber {
t.Fatalf("Expected ErrNonPositiveLinesNumber, got %s", err) t.Fatalf("Expected ErrNonPositiveLinesNumber, got %v", err)
} }
if _, err := TailFile(f, 0); err != ErrNonPositiveLinesNumber { if _, err := TailFile(f, 0); err != ErrNonPositiveLinesNumber {
t.Fatalf("Expected ErrNonPositiveLinesNumber, got %s", err) t.Fatalf("Expected ErrNonPositiveLinesNumber, got %s", err)
@ -146,3 +158,170 @@ func BenchmarkTail(b *testing.B) {
} }
} }
} }
func TestNewTailReader(t *testing.T) {
t.Parallel()
ctx := context.Background()
for dName, delim := range map[string][]byte{
"no delimiter": {},
"single byte delimiter": {'\n'},
"2 byte delimiter": []byte(";\n"),
"4 byte delimiter": []byte("####"),
"8 byte delimiter": []byte("########"),
"12 byte delimiter": []byte("############"),
} {
t.Run(dName, func(t *testing.T) {
delim := delim
t.Parallel()
s1 := "Hello world."
s2 := "Today is a fine day."
s3 := "So long, and thanks for all the fish!"
s4 := strings.Repeat("a", blockSize/2) // same as block size
s5 := strings.Repeat("a", blockSize) // just to make sure
s6 := strings.Repeat("a", blockSize*2) // bigger than block size
s7 := strings.Repeat("a", blockSize-1) // single line same as block
s8 := `{"log":"Don't panic!\n","stream":"stdout","time":"2018-04-04T20:28:44.7207062Z"}`
jsonTest := make([]string, 0, 20)
for i := 0; i < 20; i++ {
jsonTest = append(jsonTest, s8)
}
for _, test := range []struct {
desc string
data []string
}{
{desc: "one small entry", data: []string{s1}},
{desc: "several small entries", data: []string{s1, s2, s3}},
{desc: "various sizes", data: []string{s1, s2, s3, s4, s5, s1, s2, s3, s7, s6}},
{desc: "multiple lines with one more than block", data: []string{s5, s5, s5, s5, s5}},
{desc: "multiple lines much bigger than block", data: []string{s6, s6, s6, s6, s6}},
{desc: "multiple lines same as block", data: []string{s4, s4, s4, s4, s4}},
{desc: "single line same as block", data: []string{s7}},
{desc: "single line half block", data: []string{s4}},
{desc: "single line twice block", data: []string{s6}},
{desc: "json encoded values", data: jsonTest},
{desc: "no lines", data: []string{}},
{desc: "same length as delimiter", data: []string{strings.Repeat("a", len(delim))}},
} {
t.Run(test.desc, func(t *testing.T) {
test := test
t.Parallel()
max := len(test.data)
if max > 10 {
max = 10
}
s := strings.Join(test.data, string(delim))
if len(test.data) > 0 {
s += string(delim)
}
for i := 1; i <= max; i++ {
t.Run(fmt.Sprintf("%d lines", i), func(t *testing.T) {
i := i
t.Parallel()
r := strings.NewReader(s)
tr, lines, err := NewTailReaderWithDelimiter(ctx, r, i, delim)
if len(delim) == 0 {
assert.Assert(t, err != nil)
assert.Assert(t, lines == 0)
return
}
assert.Assert(t, err)
assert.Check(t, lines == i, "%d -- %d", lines, i)
b, err := ioutil.ReadAll(tr)
assert.Assert(t, err)
expectLines := test.data[len(test.data)-i:]
assert.Check(t, len(expectLines) == i)
expect := strings.Join(expectLines, string(delim)) + string(delim)
assert.Check(t, string(b) == expect, "\n%v\n%v", b, []byte(expect))
})
}
t.Run("request more lines than available", func(t *testing.T) {
t.Parallel()
r := strings.NewReader(s)
tr, lines, err := NewTailReaderWithDelimiter(ctx, r, len(test.data)*2, delim)
if len(delim) == 0 {
assert.Assert(t, err != nil)
assert.Assert(t, lines == 0)
return
}
if len(test.data) == 0 {
assert.Assert(t, err == ErrNonPositiveLinesNumber, err)
return
}
assert.Assert(t, err)
assert.Check(t, lines == len(test.data), "%d -- %d", lines, len(test.data))
b, err := ioutil.ReadAll(tr)
assert.Assert(t, err)
assert.Check(t, bytes.Equal(b, []byte(s)), "\n%v\n%v", b, []byte(s))
})
})
}
})
}
t.Run("truncated last line", func(t *testing.T) {
t.Run("more than available", func(t *testing.T) {
tail, nLines, err := NewTailReader(ctx, strings.NewReader("a\nb\nextra"), 3)
assert.Assert(t, err)
assert.Check(t, nLines == 2, nLines)
rdr := bufio.NewReader(tail)
data, _, err := rdr.ReadLine()
assert.Assert(t, err)
assert.Check(t, string(data) == "a", string(data))
data, _, err = rdr.ReadLine()
assert.Assert(t, err)
assert.Check(t, string(data) == "b", string(data))
_, _, err = rdr.ReadLine()
assert.Assert(t, err == io.EOF, err)
})
})
t.Run("truncated last line", func(t *testing.T) {
t.Run("exact", func(t *testing.T) {
tail, nLines, err := NewTailReader(ctx, strings.NewReader("a\nb\nextra"), 2)
assert.Assert(t, err)
assert.Check(t, nLines == 2, nLines)
rdr := bufio.NewReader(tail)
data, _, err := rdr.ReadLine()
assert.Assert(t, err)
assert.Check(t, string(data) == "a", string(data))
data, _, err = rdr.ReadLine()
assert.Assert(t, err)
assert.Check(t, string(data) == "b", string(data))
_, _, err = rdr.ReadLine()
assert.Assert(t, err == io.EOF, err)
})
})
t.Run("truncated last line", func(t *testing.T) {
t.Run("one line", func(t *testing.T) {
tail, nLines, err := NewTailReader(ctx, strings.NewReader("a\nb\nextra"), 1)
assert.Assert(t, err)
assert.Check(t, nLines == 1, nLines)
rdr := bufio.NewReader(tail)
data, _, err := rdr.ReadLine()
assert.Assert(t, err)
assert.Check(t, string(data) == "b", string(data))
_, _, err = rdr.ReadLine()
assert.Assert(t, err == io.EOF, err)
})
})
}