|
| 1 | +package car |
| 2 | + |
| 3 | +import ( |
| 4 | + "bytes" |
| 5 | + "context" |
| 6 | + "io" |
| 7 | + "math/rand" |
| 8 | + "testing" |
| 9 | + |
| 10 | + "github.com/ipfs/go-blockservice" |
| 11 | + "github.com/ipfs/go-cidutil" |
| 12 | + "github.com/ipfs/go-datastore" |
| 13 | + dss "github.com/ipfs/go-datastore/sync" |
| 14 | + bstore "github.com/ipfs/go-ipfs-blockstore" |
| 15 | + chunk "github.com/ipfs/go-ipfs-chunker" |
| 16 | + format "github.com/ipfs/go-ipld-format" |
| 17 | + "github.com/ipfs/go-merkledag" |
| 18 | + "github.com/ipfs/go-unixfs/importer/balanced" |
| 19 | + "github.com/ipfs/go-unixfs/importer/helpers" |
| 20 | + "github.com/ipld/go-car/v2/internal/carv1" |
| 21 | + mh "github.com/multiformats/go-multihash" |
| 22 | + "github.com/stretchr/testify/require" |
| 23 | +) |
| 24 | + |
| 25 | +func TestCarOffsetWriter(t *testing.T) { |
| 26 | + ds := dss.MutexWrap(datastore.NewMapDatastore()) |
| 27 | + bs := bstore.NewBlockstore(ds) |
| 28 | + bserv := blockservice.New(bs, nil) |
| 29 | + dserv := merkledag.NewDAGService(bserv) |
| 30 | + |
| 31 | + rseed := 5 |
| 32 | + size := 2 * 1024 * 1024 |
| 33 | + source := io.LimitReader(rand.New(rand.NewSource(int64(rseed))), int64(size)) |
| 34 | + nd, err := DAGImport(dserv, source) |
| 35 | + require.NoError(t, err) |
| 36 | + |
| 37 | + // Write the CAR to a buffer from offset 0 so the buffer can be used for |
| 38 | + // comparison |
| 39 | + payloadCid := nd.Cid() |
| 40 | + fullCarCow := NewCarOffsetWriter(payloadCid, bs) |
| 41 | + var fullBuff bytes.Buffer |
| 42 | + err = fullCarCow.Write(context.Background(), &fullBuff, 0) |
| 43 | + require.NoError(t, err) |
| 44 | + |
| 45 | + fullCar := fullBuff.Bytes() |
| 46 | + header := carHeader(nd.Cid()) |
| 47 | + headerSize, err := carv1.HeaderSize(&header) |
| 48 | + |
| 49 | + testCases := []struct { |
| 50 | + name string |
| 51 | + offset uint64 |
| 52 | + }{{ |
| 53 | + name: "1 byte offset", |
| 54 | + offset: 1, |
| 55 | + }, { |
| 56 | + name: "offset < header size", |
| 57 | + offset: headerSize - 1, |
| 58 | + }, { |
| 59 | + name: "offset == header size", |
| 60 | + offset: headerSize, |
| 61 | + }, { |
| 62 | + name: "offset > header size", |
| 63 | + offset: headerSize + 1, |
| 64 | + }, { |
| 65 | + name: "offset > header + one block size", |
| 66 | + offset: headerSize + 1024*1024 + 512*1024, |
| 67 | + }} |
| 68 | + |
| 69 | + runTestCases := func(name string, runTCWithCow func() *CarOffsetWriter) { |
| 70 | + for _, tc := range testCases { |
| 71 | + t.Run(name+" - "+tc.name, func(t *testing.T) { |
| 72 | + cow := runTCWithCow() |
| 73 | + var buff bytes.Buffer |
| 74 | + err = cow.Write(context.Background(), &buff, tc.offset) |
| 75 | + require.NoError(t, err) |
| 76 | + require.Equal(t, len(fullCar)-int(tc.offset), len(buff.Bytes())) |
| 77 | + require.Equal(t, fullCar[tc.offset:], buff.Bytes()) |
| 78 | + }) |
| 79 | + } |
| 80 | + } |
| 81 | + |
| 82 | + // Run tests with a new CarOffsetWriter |
| 83 | + runTestCases("new car offset writer", func() *CarOffsetWriter { |
| 84 | + return NewCarOffsetWriter(payloadCid, bs) |
| 85 | + }) |
| 86 | + |
| 87 | + // Run tests with a CarOffsetWriter that has already been used to write |
| 88 | + // a CAR starting at offset 0 |
| 89 | + runTestCases("fully written car offset writer", func() *CarOffsetWriter { |
| 90 | + fullCarCow := NewCarOffsetWriter(payloadCid, bs) |
| 91 | + var buff bytes.Buffer |
| 92 | + err = fullCarCow.Write(context.Background(), &buff, 0) |
| 93 | + require.NoError(t, err) |
| 94 | + return fullCarCow |
| 95 | + }) |
| 96 | + |
| 97 | + // Run tests with a CarOffsetWriter that has already been used to write |
| 98 | + // a CAR starting at offset 1 |
| 99 | + runTestCases("car offset writer written from offset 1", func() *CarOffsetWriter { |
| 100 | + fullCarCow := NewCarOffsetWriter(payloadCid, bs) |
| 101 | + var buff bytes.Buffer |
| 102 | + err = fullCarCow.Write(context.Background(), &buff, 1) |
| 103 | + require.NoError(t, err) |
| 104 | + return fullCarCow |
| 105 | + }) |
| 106 | + |
| 107 | + // Run tests with a CarOffsetWriter that has already been used to write |
| 108 | + // a CAR starting part way through the second block |
| 109 | + runTestCases("car offset writer written from offset 1.5 blocks", func() *CarOffsetWriter { |
| 110 | + fullCarCow := NewCarOffsetWriter(payloadCid, bs) |
| 111 | + var buff bytes.Buffer |
| 112 | + err = fullCarCow.Write(context.Background(), &buff, 1024*1024+512*1024) |
| 113 | + require.NoError(t, err) |
| 114 | + return fullCarCow |
| 115 | + }) |
| 116 | + |
| 117 | + // Run tests with a CarOffsetWriter that has already been used to write |
| 118 | + // a CAR repeatedly |
| 119 | + runTestCases("car offset writer written from offset repeatedly", func() *CarOffsetWriter { |
| 120 | + fullCarCow := NewCarOffsetWriter(payloadCid, bs) |
| 121 | + var buff bytes.Buffer |
| 122 | + err = fullCarCow.Write(context.Background(), &buff, 1024) |
| 123 | + require.NoError(t, err) |
| 124 | + fullCarCow = NewCarOffsetWriter(payloadCid, bs) |
| 125 | + var buff2 bytes.Buffer |
| 126 | + err = fullCarCow.Write(context.Background(), &buff2, 10) |
| 127 | + require.NoError(t, err) |
| 128 | + fullCarCow = NewCarOffsetWriter(payloadCid, bs) |
| 129 | + var buff3 bytes.Buffer |
| 130 | + err = fullCarCow.Write(context.Background(), &buff3, 1024*1024+512*1024) |
| 131 | + require.NoError(t, err) |
| 132 | + return fullCarCow |
| 133 | + }) |
| 134 | +} |
| 135 | + |
| 136 | +func TestSkipWriter(t *testing.T) { |
| 137 | + testCases := []struct { |
| 138 | + name string |
| 139 | + size int |
| 140 | + skip int |
| 141 | + expected int |
| 142 | + }{{ |
| 143 | + name: "no skip", |
| 144 | + size: 1024, |
| 145 | + skip: 0, |
| 146 | + expected: 1024, |
| 147 | + }, { |
| 148 | + name: "skip 1", |
| 149 | + size: 1024, |
| 150 | + skip: 1, |
| 151 | + expected: 1023, |
| 152 | + }, { |
| 153 | + name: "skip all", |
| 154 | + size: 1024, |
| 155 | + skip: 1024, |
| 156 | + expected: 0, |
| 157 | + }, { |
| 158 | + name: "skip overflow", |
| 159 | + size: 1024, |
| 160 | + skip: 1025, |
| 161 | + expected: 0, |
| 162 | + }} |
| 163 | + |
| 164 | + for _, tc := range testCases { |
| 165 | + t.Run(tc.name, func(t *testing.T) { |
| 166 | + var buff bytes.Buffer |
| 167 | + write := func(sw io.Writer) (int, error) { |
| 168 | + bz := make([]byte, tc.size) |
| 169 | + return sw.Write(bz) |
| 170 | + } |
| 171 | + count, err := skipWrite(&buff, uint64(tc.skip), write) |
| 172 | + require.NoError(t, err) |
| 173 | + require.Equal(t, tc.expected, count) |
| 174 | + require.Equal(t, tc.expected, len(buff.Bytes())) |
| 175 | + }) |
| 176 | + } |
| 177 | +} |
| 178 | + |
| 179 | +var DefaultHashFunction = uint64(mh.SHA2_256) |
| 180 | + |
| 181 | +func DAGImport(dserv format.DAGService, fi io.Reader) (format.Node, error) { |
| 182 | + prefix, err := merkledag.PrefixForCidVersion(1) |
| 183 | + if err != nil { |
| 184 | + return nil, err |
| 185 | + } |
| 186 | + prefix.MhType = DefaultHashFunction |
| 187 | + |
| 188 | + spl := chunk.NewSizeSplitter(fi, 1024*1024) |
| 189 | + dbp := helpers.DagBuilderParams{ |
| 190 | + Maxlinks: 1024, |
| 191 | + RawLeaves: true, |
| 192 | + |
| 193 | + CidBuilder: cidutil.InlineBuilder{ |
| 194 | + Builder: prefix, |
| 195 | + Limit: 32, |
| 196 | + }, |
| 197 | + |
| 198 | + Dagserv: dserv, |
| 199 | + } |
| 200 | + |
| 201 | + db, err := dbp.New(spl) |
| 202 | + if err != nil { |
| 203 | + return nil, err |
| 204 | + } |
| 205 | + |
| 206 | + return balanced.Layout(db) |
| 207 | +} |
0 commit comments