// repackBlob loads a single blob from src and saves it in dst. func repackBlob(src, dst *repository.Repository, id backend.ID) error { blob, err := src.Index().Lookup(id) if err != nil { return err } debug.Log("RepackBlobs", "repacking blob %v, len %v", id.Str(), blob.PlaintextLength()) buf := make([]byte, 0, blob.PlaintextLength()) buf, err = src.LoadBlob(blob.Type, id, buf) if err != nil { return err } if uint(len(buf)) != blob.PlaintextLength() { debug.Log("RepackBlobs", "repack blob %v: len(buf) isn't equal to length: %v = %v", id.Str(), len(buf), blob.PlaintextLength()) return errors.New("LoadBlob returned wrong data, len() doesn't match") } _, err = dst.SaveAndEncrypt(blob.Type, buf, &id) if err != nil { return err } return nil }
// saveRandomDataBlobs generates random data blobs and saves them to the repository. func saveRandomDataBlobs(t testing.TB, repo *repository.Repository, num int, sizeMax int) { for i := 0; i < num; i++ { size := mrand.Int() % sizeMax buf := make([]byte, size) _, err := io.ReadFull(rand.Reader, buf) OK(t, err) _, err = repo.SaveAndEncrypt(pack.Data, buf, nil) OK(t, err) } }
// saveFile reads from rd and saves the blobs in the repository. The list of // IDs is returned. func saveFile(t testing.TB, repo *repository.Repository, rd io.Reader) (blobs backend.IDs) { ch := chunker.New(rd, repo.Config.ChunkerPolynomial) for { chunk, err := ch.Next(getBuf()) if err == io.EOF { break } if err != nil { t.Fatalf("unabel to save chunk in repo: %v", err) } id, err := repo.SaveAndEncrypt(pack.Data, chunk.Data, nil) if err != nil { t.Fatalf("error saving chunk: %v", err) } blobs = append(blobs, id) } return blobs }
// ArchiveReader reads from the reader and archives the data. Returned is the // resulting snapshot and its ID. func ArchiveReader(repo *repository.Repository, p *Progress, rd io.Reader, name string) (*Snapshot, backend.ID, error) { debug.Log("ArchiveReader", "start archiving %s", name) sn, err := NewSnapshot([]string{name}) if err != nil { return nil, backend.ID{}, err } p.Start() defer p.Done() chnker := chunker.New(rd, repo.Config.ChunkerPolynomial) var ids backend.IDs var fileSize uint64 for { chunk, err := chnker.Next(getBuf()) if err == io.EOF { break } if err != nil { return nil, backend.ID{}, err } id := backend.Hash(chunk.Data) if !repo.Index().Has(id) { _, err := repo.SaveAndEncrypt(pack.Data, chunk.Data, nil) if err != nil { return nil, backend.ID{}, err } debug.Log("ArchiveReader", "saved blob %v (%d bytes)\n", id.Str(), chunk.Length) } else { debug.Log("ArchiveReader", "blob %v already saved in the repo\n", id.Str()) } freeBuf(chunk.Data) ids = append(ids, id) p.Report(Stat{Bytes: uint64(chunk.Length)}) fileSize += uint64(chunk.Length) } tree := &Tree{ Nodes: []*Node{ &Node{ Name: name, AccessTime: time.Now(), ModTime: time.Now(), Type: "file", Mode: 0644, Size: fileSize, UID: sn.UID, GID: sn.GID, User: sn.Username, Content: ids, }, }, } treeID, err := saveTreeJSON(repo, tree) if err != nil { return nil, backend.ID{}, err } sn.Tree = &treeID debug.Log("ArchiveReader", "tree saved as %v", treeID.Str()) id, err := repo.SaveJSONUnpacked(backend.Snapshot, sn) if err != nil { return nil, backend.ID{}, err } sn.id = &id debug.Log("ArchiveReader", "snapshot saved as %v", id.Str()) err = repo.Flush() if err != nil { return nil, backend.ID{}, err } err = repo.SaveIndex() if err != nil { return nil, backend.ID{}, err } return sn, id, nil }