Skip to content

Move (de)coder to its own file #487

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Mar 23, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
363 changes: 363 additions & 0 deletions codec.go
Original file line number Diff line number Diff line change
@@ -0,0 +1,363 @@
package capnp

import (
"bufio"
"encoding/binary"
"errors"
"io"
"net"

"capnproto.org/go/capnp/v3/exc"
"capnproto.org/go/capnp/v3/exp/bufferpool"
"capnproto.org/go/capnp/v3/internal/str"
"capnproto.org/go/capnp/v3/packed"
)

// A Decoder represents a framer that deserializes a particular Cap'n
// Proto input stream.
type Decoder struct {
r io.Reader

wordbuf [wordSize]byte
hdrbuf []byte

bufferPool *bufferpool.Pool

reuse bool
buf []byte
msg Message
arena roSingleSegment

// Maximum number of bytes that can be read per call to Decode.
// If not set, a reasonable default is used.
MaxMessageSize uint64
}

// NewDecoder creates a new Cap'n Proto framer that reads from r.
// The returned decoder will only read as much data as necessary to
// decode the message.
func NewDecoder(r io.Reader) *Decoder {
return &Decoder{r: r}
}

// NewPackedDecoder creates a new Cap'n Proto framer that reads from a
// packed stream r. The returned decoder may read more data than
// necessary from r.
func NewPackedDecoder(r io.Reader) *Decoder {
return NewDecoder(packed.NewReader(bufio.NewReader(r)))
}

// Decode reads a message from the decoder stream. The error is io.EOF
// only if no bytes were read.
func (d *Decoder) Decode() (*Message, error) {
maxSize := d.MaxMessageSize
if maxSize == 0 {
maxSize = defaultDecodeLimit
} else if maxSize < uint64(len(d.wordbuf)) {
return nil, errors.New("decode: max message size is smaller than header size")
}

// Read first word (number of segments and first segment size).
// For single-segment messages, this will be sufficient.
if _, err := io.ReadFull(d.r, d.wordbuf[:]); err == io.EOF {
return nil, io.EOF
} else if err != nil {
return nil, exc.WrapError("decode: read header", err)
}
maxSeg := SegmentID(binary.LittleEndian.Uint32(d.wordbuf[:]))
if maxSeg > maxStreamSegments {
return nil, errSegIDTooLarge(maxSeg)
}

// Read the rest of the header if more than one segment.
var hdr streamHeader
if maxSeg == 0 {
hdr = streamHeader{d.wordbuf[:]}
} else {
hdrSize := streamHeaderSize(maxSeg)
if hdrSize > maxSize || hdrSize > uint64(maxInt) {
return nil, errors.New("decode: message too large")
}
d.hdrbuf = resizeSlice(d.hdrbuf, int(hdrSize))
copy(d.hdrbuf, d.wordbuf[:])
if _, err := io.ReadFull(d.r, d.hdrbuf[len(d.wordbuf):]); err != nil {
return nil, exc.WrapError("decode: read header", err)
}
hdr = streamHeader{d.hdrbuf}
}
total, err := hdr.totalSize()
if err != nil {
return nil, exc.WrapError("decode", err)
}
// TODO(someday): if total size is greater than can fit in one buffer,
// attempt to allocate buffer per segment.
if total > maxSize-uint64(len(hdr.b)) || total > uint64(maxInt) {
return nil, errors.New("decode: message too large")
}

// Read segments.
if !d.reuse {
var buf []byte
if d.bufferPool == nil {
buf = make([]byte, int(total))
} else {
buf = d.bufferPool.Get(int(total))
}
if _, err := io.ReadFull(d.r, buf); err != nil {
return nil, exc.WrapError("decode: read segments", err)
}
arena, err := demuxArena(hdr, buf)
if err != nil {
return nil, exc.WrapError("decode", err)
}
return &Message{
Arena: arena,
originalBuffer: buf,
}, nil
}
d.buf = resizeSlice(d.buf, int(total))
if _, err := io.ReadFull(d.r, d.buf); err != nil {
return nil, exc.WrapError("decode: read segments", err)
}
var arena Arena
if maxSeg == 0 {
d.arena = d.buf[:len(d.buf):len(d.buf)]
arena = &d.arena
} else {
var err error
arena, err = demuxArena(hdr, d.buf)
if err != nil {
return nil, exc.WrapError("decode", err)
}
}
d.msg.Reset(arena)
return &d.msg, nil
}

type errSegIDTooLarge SegmentID

func (err errSegIDTooLarge) Error() string {
id := str.Utod(err)
max := str.Itod(maxStreamSegments)
return "decode: segment id" + id + "exceeds max segment count (max=" + max + ")"
}

func resizeSlice(b []byte, size int) []byte {
if cap(b) < size {
return make([]byte, size)
}
return b[:size]
}

// ReuseBuffer causes the decoder to reuse its buffer on subsequent decodes.
// The decoder may return messages that cannot handle allocations.
func (d *Decoder) ReuseBuffer() {
d.reuse = true
}

// SetBufferPool registers a buffer pool to allocate message space from, rather
// than directly allocating buffers with make(). This can help reduce pressure
// on the garbage collector; pass messages to d.ReleaseMessage() when done with
// them.
func (d *Decoder) SetBufferPool(p *bufferpool.Pool) {
d.bufferPool = p
}

func (d *Decoder) ReleaseMessage(m *Message) {
if d.bufferPool == nil {
return
}
d.bufferPool.Put(m.originalBuffer)
}

// Unmarshal reads an unpacked serialized stream into a message. No
// copying is performed, so the objects in the returned message read
// directly from data.
func Unmarshal(data []byte) (*Message, error) {
if len(data) == 0 {
return nil, io.EOF
}
if len(data) < int(wordSize) {
return nil, errors.New("unmarshal: short header section")
}
maxSeg := SegmentID(binary.LittleEndian.Uint32(data))
hdrSize := streamHeaderSize(maxSeg)
if uint64(len(data)) < hdrSize {
return nil, errors.New("unmarshal: short header section")
}
hdr := streamHeader{data[:hdrSize]}
data = data[hdrSize:]
if total, err := hdr.totalSize(); err != nil {
return nil, exc.WrapError("unmarshal", err)
} else if total > uint64(len(data)) {
return nil, errors.New("unmarshal: short data section")
}
arena, err := demuxArena(hdr, data)
if err != nil {
return nil, exc.WrapError("unmarshal", err)
}
return &Message{Arena: arena}, nil
}

// UnmarshalPacked reads a packed serialized stream into a message.
func UnmarshalPacked(data []byte) (*Message, error) {
if len(data) == 0 {
return nil, io.EOF
}
data, err := packed.Unpack(nil, data)
if err != nil {
return nil, exc.WrapError("unmarshal", err)
}
return Unmarshal(data)
}

// MustUnmarshalRoot reads an unpacked serialized stream and returns
// its root pointer. If there is any error, it panics.
func MustUnmarshalRoot(data []byte) Ptr {
msg, err := Unmarshal(data)
if err != nil {
panic(err)
}
p, err := msg.Root()
if err != nil {
panic(err)
}
return p
}

// An Encoder represents a framer for serializing a particular Cap'n
// Proto stream.
type Encoder struct {
w io.Writer
hdrbuf []byte
bufs [][]byte
}

// NewEncoder creates a new Cap'n Proto framer that writes to w.
func NewEncoder(w io.Writer) *Encoder {
return &Encoder{w: w}
}

// NewPackedEncoder creates a new Cap'n Proto framer that writes to a
// packed stream w.
func NewPackedEncoder(w io.Writer) *Encoder {
return NewEncoder(&packed.Writer{Writer: w})
}

// Encode writes a message to the encoder stream.
func (e *Encoder) Encode(m *Message) error {
nsegs := m.NumSegments()
if nsegs == 0 {
return errors.New("encode: message has no segments")
}
e.bufs = append(e.bufs[:0], nil) // first element is placeholder for header
maxSeg := SegmentID(nsegs - 1)
hdrSize := streamHeaderSize(maxSeg)
if hdrSize > uint64(maxInt) {
return errors.New("encode: header size overflows int")
}
e.hdrbuf = resizeSlice(e.hdrbuf, int(hdrSize))
e.hdrbuf = appendUint32(e.hdrbuf[:0], uint32(maxSeg))
for i := int64(0); i < nsegs; i++ {
s, err := m.Segment(SegmentID(i))
if err != nil {
return exc.WrapError("encode", err)
}
n := len(s.data)
if int64(n) > int64(maxSegmentSize) {
return errors.New("encode: segment " + str.Itod(i) + " too large")
}
e.hdrbuf = appendUint32(e.hdrbuf, uint32(Size(n)/wordSize))
e.bufs = append(e.bufs, s.data)
}
if len(e.hdrbuf)%int(wordSize) != 0 {
e.hdrbuf = appendUint32(e.hdrbuf, 0)
}
e.bufs[0] = e.hdrbuf

if err := e.write(e.bufs); err != nil {
return exc.WrapError("encode", err)
}

return nil
}

func (e *Encoder) write(bufs [][]byte) error {
_, err := (*net.Buffers)(&bufs).WriteTo(e.w)
return err
}

// streamHeaderSize returns the size of the header, given the lower 32
// bits of the first word of the header (the number of segments minus
// one).
func streamHeaderSize(maxSeg SegmentID) uint64 {
return ((uint64(maxSeg)+2)*4 + 7) &^ 7
}

// appendUint32 appends a uint32 to a byte slice and returns the
// new slice.
func appendUint32(b []byte, v uint32) []byte {
b = append(b, 0, 0, 0, 0)
binary.LittleEndian.PutUint32(b[len(b)-4:], v)
return b
}

// streamHeader holds the framing words at the beginning of a serialized
// Cap'n Proto message.
type streamHeader struct {
b []byte
}

// maxSegment returns the number of segments in the message minus one.
func (h streamHeader) maxSegment() SegmentID {
return SegmentID(binary.LittleEndian.Uint32(h.b))
}

// segmentSize returns the size of segment i, returning an error if the
// segment overflows maxSegmentSize.
func (h streamHeader) segmentSize(i SegmentID) (Size, error) {
s := binary.LittleEndian.Uint32(h.b[4+i*4:])
sz, ok := wordSize.times(int32(s))
if !ok {
return 0, errors.New("segment " + str.Utod(i) + ": overflow size")
}
return sz, nil
}

// totalSize returns the sum of all the segment sizes. The sum will
// be in the range [0, 0xfffffff800000000].
func (h streamHeader) totalSize() (uint64, error) {
var sum uint64
for i := uint64(0); i <= uint64(h.maxSegment()); i++ {
x, err := h.segmentSize(SegmentID(i))
if err != nil {
return sum, err
}
sum += uint64(x)
}
return sum, nil
}

func hasCapacity(b []byte, sz Size) bool {
return sz <= Size(cap(b)-len(b))
}

// demuxArena slices b into a multi-segment arena. It assumes that
// len(data) >= hdr.totalSize().
func demuxArena(hdr streamHeader, data []byte) (Arena, error) {
maxSeg := hdr.maxSegment()
if int64(maxSeg) > int64(maxInt-1) {
return nil, errors.New("number of segments overflows int")
}
segs := make([][]byte, int(maxSeg+1))
for i := range segs {
sz, err := hdr.segmentSize(SegmentID(i))
if err != nil {
return nil, err
}
segs[i], data = data[:sz:sz], data[sz:]
}
return MultiSegment(segs), nil
}

const maxInt = int(^uint(0) >> 1)
Loading