1
0
mirror of https://github.com/taigrr/bitcask synced 2025-01-18 04:03:17 -08:00

fix: check of persisted index values (#91)

This commit is contained in:
Ignacio Hagopian 2019-09-04 09:42:32 -03:00 committed by James Mills
parent 1c7df7f9c7
commit a2b5ae2287
3 changed files with 13 additions and 20 deletions

View File

@ -337,7 +337,7 @@ func (b *Bitcask) reopen() error {
} }
defer f.Close() defer f.Close()
if err := internal.ReadIndex(f, t, b.config.maxKeySize, b.config.maxValueSize); err != nil { if err := internal.ReadIndex(f, t, b.config.maxKeySize); err != nil {
return err return err
} }
} else { } else {

View File

@ -13,7 +13,6 @@ var (
errTruncatedKeyData = errors.New("key data is truncated") errTruncatedKeyData = errors.New("key data is truncated")
errTruncatedData = errors.New("data is truncated") errTruncatedData = errors.New("data is truncated")
errKeySizeTooLarge = errors.New("key size too large") errKeySizeTooLarge = errors.New("key size too large")
errDataSizeTooLarge = errors.New("data size too large")
) )
const ( const (
@ -60,21 +59,17 @@ func writeBytes(b []byte, w io.Writer) error {
return nil return nil
} }
func readItem(r io.Reader, maxValueSize int) (Item, error) { func readItem(r io.Reader) (Item, error) {
buf := make([]byte, (fileIDSize + offsetSize + sizeSize)) buf := make([]byte, (fileIDSize + offsetSize + sizeSize))
_, err := io.ReadFull(r, buf) _, err := io.ReadFull(r, buf)
if err != nil { if err != nil {
return Item{}, errors.Wrap(errTruncatedData, err.Error()) return Item{}, errors.Wrap(errTruncatedData, err.Error())
} }
size := int64(binary.BigEndian.Uint64(buf[(fileIDSize + offsetSize):]))
if size > int64(maxValueSize) {
return Item{}, errDataSizeTooLarge
}
return Item{ return Item{
FileID: int(binary.BigEndian.Uint32(buf[:fileIDSize])), FileID: int(binary.BigEndian.Uint32(buf[:fileIDSize])),
Offset: int64(binary.BigEndian.Uint64(buf[fileIDSize:(fileIDSize + offsetSize)])), Offset: int64(binary.BigEndian.Uint64(buf[fileIDSize:(fileIDSize + offsetSize)])),
Size: size, Size: int64(binary.BigEndian.Uint64(buf[(fileIDSize + offsetSize):])),
}, nil }, nil
} }
@ -91,7 +86,7 @@ func writeItem(item Item, w io.Writer) error {
} }
// ReadIndex reads a persisted from a io.Reader into a Tree // ReadIndex reads a persisted from a io.Reader into a Tree
func ReadIndex(r io.Reader, t art.Tree, maxKeySize, maxValueSize int) error { func ReadIndex(r io.Reader, t art.Tree, maxKeySize int) error {
for { for {
key, err := readKeyBytes(r, maxKeySize) key, err := readKeyBytes(r, maxKeySize)
if err != nil { if err != nil {
@ -101,7 +96,7 @@ func ReadIndex(r io.Reader, t art.Tree, maxKeySize, maxValueSize int) error {
return err return err
} }
item, err := readItem(r, maxValueSize) item, err := readItem(r)
if err != nil { if err != nil {
return err return err
} }

View File

@ -36,7 +36,7 @@ func TestReadIndex(t *testing.T) {
b := bytes.NewBuffer(sampleTreeBytes) b := bytes.NewBuffer(sampleTreeBytes)
at := art.New() at := art.New()
err := ReadIndex(b, at, 1024, 1024) err := ReadIndex(b, at, 1024)
if err != nil { if err != nil {
t.Fatalf("error while deserializing correct sample tree: %v", err) t.Fatalf("error while deserializing correct sample tree: %v", err)
} }
@ -74,7 +74,7 @@ func TestReadCorruptedData(t *testing.T) {
t.Run(table[i].name, func(t *testing.T) { t.Run(table[i].name, func(t *testing.T) {
bf := bytes.NewBuffer(table[i].data) bf := bytes.NewBuffer(table[i].data)
if err := ReadIndex(bf, art.New(), 1024, 1024); errors.Cause(err) != table[i].err { if err := ReadIndex(bf, art.New(), 1024); errors.Cause(err) != table[i].err {
t.Fatalf("expected %v, got %v", table[i].err, err) t.Fatalf("expected %v, got %v", table[i].err, err)
} }
}) })
@ -94,18 +94,16 @@ func TestReadCorruptedData(t *testing.T) {
name string name string
err error err error
maxKeySize int maxKeySize int
maxValueSize int
data []byte data []byte
}{ }{
{name: "key-data-overflow", err: errKeySizeTooLarge, maxKeySize: 1024, maxValueSize: 1024, data: overflowKeySize}, {name: "key-data-overflow", err: errKeySizeTooLarge, maxKeySize: 1024, data: overflowKeySize},
{name: "item-data-overflow", err: errDataSizeTooLarge, maxKeySize: 1024, maxValueSize: 1024, data: overflowDataSize},
} }
for i := range table { for i := range table {
t.Run(table[i].name, func(t *testing.T) { t.Run(table[i].name, func(t *testing.T) {
bf := bytes.NewBuffer(table[i].data) bf := bytes.NewBuffer(table[i].data)
if err := ReadIndex(bf, art.New(), table[i].maxKeySize, table[i].maxValueSize); errors.Cause(err) != table[i].err { if err := ReadIndex(bf, art.New(), table[i].maxKeySize); errors.Cause(err) != table[i].err {
t.Fatalf("expected %v, got %v", table[i].err, err) t.Fatalf("expected %v, got %v", table[i].err, err)
} }
}) })