mirror of
https://github.com/taigrr/bitcask
synced 2025-01-18 04:03:17 -08:00
fix: check of persisted index values (#91)
This commit is contained in:
parent
1c7df7f9c7
commit
a2b5ae2287
@ -337,7 +337,7 @@ func (b *Bitcask) reopen() error {
|
|||||||
}
|
}
|
||||||
defer f.Close()
|
defer f.Close()
|
||||||
|
|
||||||
if err := internal.ReadIndex(f, t, b.config.maxKeySize, b.config.maxValueSize); err != nil {
|
if err := internal.ReadIndex(f, t, b.config.maxKeySize); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
|
@ -13,7 +13,6 @@ var (
|
|||||||
errTruncatedKeyData = errors.New("key data is truncated")
|
errTruncatedKeyData = errors.New("key data is truncated")
|
||||||
errTruncatedData = errors.New("data is truncated")
|
errTruncatedData = errors.New("data is truncated")
|
||||||
errKeySizeTooLarge = errors.New("key size too large")
|
errKeySizeTooLarge = errors.New("key size too large")
|
||||||
errDataSizeTooLarge = errors.New("data size too large")
|
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@ -60,21 +59,17 @@ func writeBytes(b []byte, w io.Writer) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func readItem(r io.Reader, maxValueSize int) (Item, error) {
|
func readItem(r io.Reader) (Item, error) {
|
||||||
buf := make([]byte, (fileIDSize + offsetSize + sizeSize))
|
buf := make([]byte, (fileIDSize + offsetSize + sizeSize))
|
||||||
_, err := io.ReadFull(r, buf)
|
_, err := io.ReadFull(r, buf)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return Item{}, errors.Wrap(errTruncatedData, err.Error())
|
return Item{}, errors.Wrap(errTruncatedData, err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
size := int64(binary.BigEndian.Uint64(buf[(fileIDSize + offsetSize):]))
|
|
||||||
if size > int64(maxValueSize) {
|
|
||||||
return Item{}, errDataSizeTooLarge
|
|
||||||
}
|
|
||||||
return Item{
|
return Item{
|
||||||
FileID: int(binary.BigEndian.Uint32(buf[:fileIDSize])),
|
FileID: int(binary.BigEndian.Uint32(buf[:fileIDSize])),
|
||||||
Offset: int64(binary.BigEndian.Uint64(buf[fileIDSize:(fileIDSize + offsetSize)])),
|
Offset: int64(binary.BigEndian.Uint64(buf[fileIDSize:(fileIDSize + offsetSize)])),
|
||||||
Size: size,
|
Size: int64(binary.BigEndian.Uint64(buf[(fileIDSize + offsetSize):])),
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -91,7 +86,7 @@ func writeItem(item Item, w io.Writer) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// ReadIndex reads a persisted from a io.Reader into a Tree
|
// ReadIndex reads a persisted from a io.Reader into a Tree
|
||||||
func ReadIndex(r io.Reader, t art.Tree, maxKeySize, maxValueSize int) error {
|
func ReadIndex(r io.Reader, t art.Tree, maxKeySize int) error {
|
||||||
for {
|
for {
|
||||||
key, err := readKeyBytes(r, maxKeySize)
|
key, err := readKeyBytes(r, maxKeySize)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -101,7 +96,7 @@ func ReadIndex(r io.Reader, t art.Tree, maxKeySize, maxValueSize int) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
item, err := readItem(r, maxValueSize)
|
item, err := readItem(r)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
@ -36,7 +36,7 @@ func TestReadIndex(t *testing.T) {
|
|||||||
b := bytes.NewBuffer(sampleTreeBytes)
|
b := bytes.NewBuffer(sampleTreeBytes)
|
||||||
|
|
||||||
at := art.New()
|
at := art.New()
|
||||||
err := ReadIndex(b, at, 1024, 1024)
|
err := ReadIndex(b, at, 1024)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("error while deserializing correct sample tree: %v", err)
|
t.Fatalf("error while deserializing correct sample tree: %v", err)
|
||||||
}
|
}
|
||||||
@ -74,7 +74,7 @@ func TestReadCorruptedData(t *testing.T) {
|
|||||||
t.Run(table[i].name, func(t *testing.T) {
|
t.Run(table[i].name, func(t *testing.T) {
|
||||||
bf := bytes.NewBuffer(table[i].data)
|
bf := bytes.NewBuffer(table[i].data)
|
||||||
|
|
||||||
if err := ReadIndex(bf, art.New(), 1024, 1024); errors.Cause(err) != table[i].err {
|
if err := ReadIndex(bf, art.New(), 1024); errors.Cause(err) != table[i].err {
|
||||||
t.Fatalf("expected %v, got %v", table[i].err, err)
|
t.Fatalf("expected %v, got %v", table[i].err, err)
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
@ -91,21 +91,19 @@ func TestReadCorruptedData(t *testing.T) {
|
|||||||
binary.BigEndian.PutUint32(overflowDataSize[int32Size+4+fileIDSize+offsetSize:], 1025)
|
binary.BigEndian.PutUint32(overflowDataSize[int32Size+4+fileIDSize+offsetSize:], 1025)
|
||||||
|
|
||||||
table := []struct {
|
table := []struct {
|
||||||
name string
|
name string
|
||||||
err error
|
err error
|
||||||
maxKeySize int
|
maxKeySize int
|
||||||
maxValueSize int
|
data []byte
|
||||||
data []byte
|
|
||||||
}{
|
}{
|
||||||
{name: "key-data-overflow", err: errKeySizeTooLarge, maxKeySize: 1024, maxValueSize: 1024, data: overflowKeySize},
|
{name: "key-data-overflow", err: errKeySizeTooLarge, maxKeySize: 1024, data: overflowKeySize},
|
||||||
{name: "item-data-overflow", err: errDataSizeTooLarge, maxKeySize: 1024, maxValueSize: 1024, data: overflowDataSize},
|
|
||||||
}
|
}
|
||||||
|
|
||||||
for i := range table {
|
for i := range table {
|
||||||
t.Run(table[i].name, func(t *testing.T) {
|
t.Run(table[i].name, func(t *testing.T) {
|
||||||
bf := bytes.NewBuffer(table[i].data)
|
bf := bytes.NewBuffer(table[i].data)
|
||||||
|
|
||||||
if err := ReadIndex(bf, art.New(), table[i].maxKeySize, table[i].maxValueSize); errors.Cause(err) != table[i].err {
|
if err := ReadIndex(bf, art.New(), table[i].maxKeySize); errors.Cause(err) != table[i].err {
|
||||||
t.Fatalf("expected %v, got %v", table[i].err, err)
|
t.Fatalf("expected %v, got %v", table[i].err, err)
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
Loading…
x
Reference in New Issue
Block a user