func (vs *VolumeServer) getVolumeDataContentHandler(w http.ResponseWriter, r *http.Request) { v, err := vs.getVolume("volume", r) if v == nil { writeJsonError(w, r, http.StatusBadRequest, fmt.Errorf("Not Found volume: %v", err)) return } if int(v.SuperBlock.CompactRevision) != util.ParseInt(r.FormValue("revision"), 0) { writeJsonError(w, r, http.StatusExpectationFailed, fmt.Errorf("Requested Volume Revision is %s, but current revision is %d", r.FormValue("revision"), v.SuperBlock.CompactRevision)) return } offset := uint32(util.ParseUint64(r.FormValue("offset"), 0)) size := uint32(util.ParseUint64(r.FormValue("size"), 0)) content, block, err := storage.ReadNeedleBlob(v.DataFile(), int64(offset)*storage.NeedlePaddingSize, size) defer storage.ReleaseBytes(block.Bytes) if err != nil { writeJsonError(w, r, http.StatusInternalServerError, err) return } id := util.ParseUint64(r.FormValue("id"), 0) n := new(storage.Needle) n.ParseNeedleHeader(content) if id != n.Id { writeJsonError(w, r, http.StatusNotFound, fmt.Errorf("Expected file entry id %d, but found %d", id, n.Id)) return } w.Write(content) }
func (vs *VolumeServer) tryHandleChunkedFile(n *storage.Needle, fileName string, w http.ResponseWriter, r *http.Request) (processed bool) { if !n.IsChunkedManifest() { return false } chunkManifest, e := operation.LoadChunkManifest(n.Data, n.IsGzipped()) if e != nil { glog.V(0).Infof("load chunked manifest (%s) error: %v", r.URL.Path, e) return false } if fileName == "" && chunkManifest.Name != "" { fileName = chunkManifest.Name } mType := "" if chunkManifest.Mime != "" { mt := chunkManifest.Mime if !strings.HasPrefix(mt, "application/octet-stream") { mType = mt } } w.Header().Set("X-File-Store", "chunked") chunkedFileReader := &operation.ChunkedFileReader{ Manifest: chunkManifest, Master: vs.GetMasterNode(), } defer chunkedFileReader.Close() if e := writeResponseContent(fileName, mType, chunkedFileReader, w, r); e != nil { glog.V(2).Infoln("response write error:", e) } return true }
func ReplicatedWrite(masterNode string, s *storage.Store, volumeId storage.VolumeId, needle *storage.Needle, r *http.Request) (size uint32, errorStatus string) { //check JWT jwt := security.GetJwt(r) ret, err := s.Write(volumeId, needle) needToReplicate := !s.HasVolume(volumeId) if err != nil { errorStatus = "Failed to write to local disk (" + err.Error() + ")" } else if ret > 0 { needToReplicate = needToReplicate || s.GetVolume(volumeId).NeedToReplicate() } else { errorStatus = "Failed to write to local disk" } if !needToReplicate && ret > 0 { needToReplicate = s.GetVolume(volumeId).NeedToReplicate() } if needToReplicate { //send to other replica locations if r.FormValue("type") != "replicate" { if err = distributedOperation(masterNode, s, volumeId, func(location operation.Location) error { u := url.URL{ Scheme: "http", Host: location.Url, Path: r.URL.Path, } q := url.Values{ "type": {"replicate"}, } if needle.LastModified > 0 { q.Set("ts", strconv.FormatUint(needle.LastModified, 10)) } if needle.IsChunkedManifest() { q.Set("cm", "true") } u.RawQuery = q.Encode() _, err := operation.Upload(u.String(), string(needle.Name), bytes.NewReader(needle.Data), needle.IsGzipped(), string(needle.Mime), jwt) return err }); err != nil { ret = 0 errorStatus = fmt.Sprintf("Failed to write to replicas for volume %d: %v", volumeId, err) } } } size = ret return }
func ReplicatedWrite(masterNode string, s *storage.Store, volumeId storage.VolumeId, needle *storage.Needle, r *http.Request) (size uint32, errorStatus string) { //check JWT jwt := security.GetJwt(r) defer func() { if errorStatus == "" { return } ReplicatedDelete(masterNode, s, volumeId, needle, r) }() ret, err := s.Write(volumeId, needle) if err != nil { errorStatus = "Failed to write to local disk (" + err.Error() + ")" } else if ret <= 0 { errorStatus = "Failed to write to local disk" } //send to other replica locations if r.FormValue("type") != "replicate" { repWrite := func(location operation.Location) bool { args := url.Values{ "type": {"replicate"}, } if needle.LastModified > 0 { args.Set("ts", strconv.FormatUint(needle.LastModified, 10)) } if needle.IsChunkedManifest() { args.Set("cm", "true") } u := util.MkUrl(location.Url, r.URL.Path, args) glog.V(4).Infoln("write replication to", u) _, err := operation.Upload(u, string(needle.Name), bytes.NewReader(needle.Data), needle.IsGzipped(), string(needle.Mime), jwt) if err != nil { glog.V(0).Infof("write replication to %s err, %v", u, err) } return err == nil } if !distributedOperation(masterNode, s, volumeId, repWrite) { ret = 0 errorStatus = "Failed to write to replicas for volume " + volumeId.String() } } size = ret return }
func (vs *VolumeServer) DeleteHandler(w http.ResponseWriter, r *http.Request) { n := new(storage.Needle) vid, fid, _, _, _ := parseURLPath(r.URL.Path) volumeId, _ := storage.NewVolumeId(vid) n.ParsePath(fid) glog.V(2).Infoln("deleting", n) cookie := n.Cookie _, ok := vs.store.ReadVolumeNeedle(volumeId, n) if ok != nil { m := make(map[string]uint32) m["size"] = 0 writeJsonQuiet(w, r, http.StatusNotFound, m) return } defer n.ReleaseMemory() if n.Cookie != cookie { glog.V(0).Infoln("delete", r.URL.Path, "with unmaching cookie from ", r.RemoteAddr, "agent", r.UserAgent()) writeJsonError(w, r, http.StatusBadRequest, errors.New("File Random Cookie does not match.")) return } count := int64(n.Size) if n.IsChunkedManifest() { chunkManifest, e := operation.LoadChunkManifest(n.Data, n.IsGzipped()) if e != nil { writeJsonError(w, r, http.StatusInternalServerError, fmt.Errorf("Load chunks manifest error: %v", e)) return } // make sure all chunks had deleted before delete manifest if e := chunkManifest.DeleteChunks(vs.GetMasterNode()); e != nil { writeJsonError(w, r, http.StatusInternalServerError, fmt.Errorf("Delete chunks error: %v", e)) return } count = chunkManifest.Size } _, err := topology.ReplicatedDelete(vs.GetMasterNode(), vs.store, volumeId, n, r) if err == nil { m := make(map[string]int64) m["size"] = count writeJsonQuiet(w, r, http.StatusAccepted, m) } else { writeJsonError(w, r, http.StatusInternalServerError, fmt.Errorf("Deletion Failed: %v", err)) } }
func walker(vid storage.VolumeId, n *storage.Needle, version storage.Version) (err error) { key := storage.NewFileIdFromNeedle(vid, n).String() if tarOutputFile != nil { fileNameTemplateBuffer.Reset() if err = fileNameTemplate.Execute(fileNameTemplateBuffer, nameParams{ Name: string(n.Name), Id: n.Id, Mime: string(n.Mime), Key: key, Ext: filepath.Ext(string(n.Name)), }, ); err != nil { return err } fileName := fileNameTemplateBuffer.String() if n.IsGzipped() && path.Ext(fileName) != ".gz" { fileName = fileName + ".gz" } tarHeader.Name, tarHeader.Size = fileName, int64(len(n.Data)) if n.HasLastModifiedDate() { tarHeader.ModTime = time.Unix(int64(n.LastModified), 0) } else { tarHeader.ModTime = time.Unix(0, 0) } tarHeader.ChangeTime = tarHeader.ModTime if err = tarOutputFile.WriteHeader(&tarHeader); err != nil { return err } _, err = tarOutputFile.Write(n.Data) } else { size := n.DataSize if version == storage.Version1 { size = n.Size } fmt.Printf("key=%s Name=%s Size=%d gzip=%t mime=%s\n", key, n.Name, size, n.IsGzipped(), n.Mime, ) } return }
//Experts only: takes multiple fid parameters. This function does not propagate deletes to replicas. func (vs *VolumeServer) batchDeleteHandler(w http.ResponseWriter, r *http.Request) { r.ParseForm() var ret []operation.DeleteResult for _, fid := range r.Form["fid"] { vid, id_cookie, err := operation.ParseFileId(fid) if err != nil { ret = append(ret, operation.DeleteResult{ Fid: fid, Status: http.StatusBadRequest, Error: err.Error()}) continue } n := new(storage.Needle) volumeId, _ := storage.NewVolumeId(vid) n.ParsePath(id_cookie) glog.V(4).Infoln("batch deleting", n) cookie := n.Cookie if _, err := vs.store.ReadVolumeNeedle(volumeId, n); err != nil { ret = append(ret, operation.DeleteResult{ Fid: fid, Status: http.StatusNotFound, Error: err.Error(), }) continue } if n.IsChunkedManifest() { ret = append(ret, operation.DeleteResult{ Fid: fid, Status: http.StatusNotAcceptable, Error: "ChunkManifest: not allowed in batch delete mode.", }) n.ReleaseMemory() continue } if n.Cookie != cookie { ret = append(ret, operation.DeleteResult{ Fid: fid, Status: http.StatusBadRequest, Error: "File Random Cookie does not match.", }) glog.V(0).Infoln("deleting", fid, "with unmaching cookie from ", r.RemoteAddr, "agent", r.UserAgent()) n.ReleaseMemory() return } if size, err := vs.store.Delete(volumeId, n); err != nil { ret = append(ret, operation.DeleteResult{ Fid: fid, Status: http.StatusInternalServerError, Error: err.Error()}, ) } else { ret = append(ret, operation.DeleteResult{ Fid: fid, Status: http.StatusAccepted, Size: int(size)}, ) } n.ReleaseMemory() } writeJsonQuiet(w, r, http.StatusAccepted, ret) }
func (vs *VolumeServer) GetOrHeadHandler(w http.ResponseWriter, r *http.Request) { vid, nid, filename, ext, _ := parseURLPath(r.URL.Path) volumeId, err := storage.NewVolumeId(vid) if err != nil { glog.V(2).Infoln("parsing error:", err, r.URL.Path) w.WriteHeader(http.StatusBadRequest) return } var ( n *storage.Needle ) fid, err := storage.NewFileIdFromNid(vid, nid) if err != nil { glog.V(2).Infoln("parsing fid error:", err, r.URL.Path) w.WriteHeader(http.StatusBadRequest) return } glog.V(4).Infoln("volume", volumeId, "reading", n) if vs.store.HasVolume(volumeId) { n, err = vs.store.ReadLocalNeedle(fid) glog.V(4).Infoln("read local needle", fid, "error", err) if err != nil { glog.V(0).Infoln("read local error:", err, r.URL.Path) w.WriteHeader(http.StatusNotFound) return } } else if vs.ReadRemoteNeedle { n, err = vs.store.ReadRemoteNeedle(fid, r.FormValue("collection")) glog.V(4).Infoln("read remote needle", fid, "error", err) if err != nil { glog.V(0).Infoln("read remote error:", err, ",url path:", r.URL.Path) w.WriteHeader(http.StatusNotFound) return } } else if vs.ReadRedirect { lookupResult, err := operation.Lookup(vs.GetMasterNode(), volumeId.String(), r.FormValue("collection")) glog.V(2).Infoln("volume", volumeId, "found on", lookupResult, "error", err) if err == nil && len(lookupResult.Locations) > 0 { u, _ := url.Parse(util.NormalizeUrl(lookupResult.Locations.PickForRead().PublicUrl)) u.Path = r.URL.Path http.Redirect(w, r, u.String(), http.StatusMovedPermanently) } else { glog.V(2).Infoln("lookup error:", err, r.URL.Path) w.WriteHeader(http.StatusNotFound) } return } else { glog.V(2).Infoln("volume is not local:", err, r.URL.Path) w.WriteHeader(http.StatusNotFound) return } if n.LastModified != 0 { w.Header().Set("Last-Modified", time.Unix(int64(n.LastModified), 0).UTC().Format(http.TimeFormat)) if r.Header.Get("If-Modified-Since") != "" { if t, parseError := time.Parse(http.TimeFormat, r.Header.Get("If-Modified-Since")); parseError == nil { if t.Unix() >= int64(n.LastModified) { w.WriteHeader(http.StatusNotModified) return } } } } etag := n.Etag() if inm := r.Header.Get("If-None-Match"); inm == etag { w.WriteHeader(http.StatusNotModified) return } w.Header().Set("Etag", etag) if vs.tryHandleChunkedFile(volumeId, n, filename, w, r) { return } if n.NameSize > 0 && filename == "" { filename = string(n.Name) if ext == "" { ext = path.Ext(filename) } } mtype := "" if n.MimeSize > 0 { mt := string(n.Mime) if !strings.HasPrefix(mt, "application/octet-stream") { mtype = mt } } needleData := n.Data if ext != ".gz" { if n.IsGzipped() { if strings.Contains(r.Header.Get("Accept-Encoding"), "gzip") { w.Header().Set("Content-Encoding", "gzip") } else { if needleData, err = operation.UnGzipData(needleData); err != nil { glog.V(0).Infoln("ungzip error:", err, r.URL.Path) } } } } if ext == ".png" || ext == ".jpg" || ext == ".gif" { width, height := 0, 0 if r.FormValue("width") != "" { width, _ = strconv.Atoi(r.FormValue("width")) } if r.FormValue("height") != "" { height, _ = strconv.Atoi(r.FormValue("height")) } if needleData, _, _, err = images.Resized(ext, needleData, width, height); err != nil { glog.V(0).Infoln("resize image error,", err, r.URL.Path) } } if err = writeResponseContent(filename, mtype, bytes.NewReader(needleData), w, r); err != nil { glog.V(2).Infoln("response write error:", err, r.URL.Path) } }
func (vs *VolumeServer) GetOrHeadHandler(w http.ResponseWriter, r *http.Request) { n := new(storage.Needle) vid, fid, filename, ext, _ := parseURLPath(r.URL.Path) volumeId, err := storage.NewVolumeId(vid) if err != nil { glog.V(2).Infoln("parsing error:", err, r.URL.Path) w.WriteHeader(http.StatusBadRequest) return } err = n.ParsePath(fid) if err != nil { glog.V(2).Infoln("parsing fid error:", err, r.URL.Path) w.WriteHeader(http.StatusBadRequest) return } glog.V(4).Infoln("volume", volumeId, "reading", n) if !vs.store.HasVolume(volumeId) { if !vs.ReadRedirect { glog.V(2).Infoln("volume is not local:", err, r.URL.Path) w.WriteHeader(http.StatusNotFound) return } lookupResult, err := operation.Lookup(vs.GetMasterNode(), volumeId.String()) glog.V(2).Infoln("volume", volumeId, "found on", lookupResult, "error", err) if err == nil && len(lookupResult.Locations) > 0 { u, _ := url.Parse(util.NormalizeUrl(lookupResult.Locations[0].PublicUrl)) u.Path = r.URL.Path arg := url.Values{} if c := r.FormValue("collection"); c != "" { arg.Set("collection", c) } u.RawQuery = arg.Encode() http.Redirect(w, r, u.String(), http.StatusMovedPermanently) } else { glog.V(2).Infoln("lookup error:", err, r.URL.Path) w.WriteHeader(http.StatusNotFound) } return } cookie := n.Cookie count, e := vs.store.ReadVolumeNeedle(volumeId, n) glog.V(4).Infoln("read bytes", count, "error", e) if e != nil || count <= 0 { glog.V(0).Infoln("read error:", e, r.URL.Path) w.WriteHeader(http.StatusNotFound) return } defer n.ReleaseMemory() if n.Cookie != cookie { glog.V(0).Infoln("request", r.URL.Path, "with unmaching cookie seen:", cookie, "expected:", n.Cookie, "from", r.RemoteAddr, "agent", r.UserAgent()) w.WriteHeader(http.StatusNotFound) return } if n.LastModified != 0 { w.Header().Set("Last-Modified", time.Unix(int64(n.LastModified), 0).UTC().Format(http.TimeFormat)) if r.Header.Get("If-Modified-Since") != "" { if t, parseError := time.Parse(http.TimeFormat, r.Header.Get("If-Modified-Since")); parseError == nil { if t.Unix() >= int64(n.LastModified) { w.WriteHeader(http.StatusNotModified) return } } } } etag := n.Etag() if inm := r.Header.Get("If-None-Match"); inm == etag { w.WriteHeader(http.StatusNotModified) return } w.Header().Set("Etag", etag) if vs.tryHandleChunkedFile(n, filename, w, r) { return } if n.NameSize > 0 && filename == "" { filename = string(n.Name) if ext == "" { ext = path.Ext(filename) } } mtype := "" if n.MimeSize > 0 { mt := string(n.Mime) if !strings.HasPrefix(mt, "application/octet-stream") { mtype = mt } } if ext != ".gz" { if n.IsGzipped() { if strings.Contains(r.Header.Get("Accept-Encoding"), "gzip") { w.Header().Set("Content-Encoding", "gzip") } else { if n.Data, err = operation.UnGzipData(n.Data); err != nil { glog.V(0).Infoln("ungzip error:", err, r.URL.Path) } } } } if ext == ".png" || ext == ".jpg" || ext == ".gif" { width, height := 0, 0 if r.FormValue("width") != "" { width, _ = strconv.Atoi(r.FormValue("width")) } if r.FormValue("height") != "" { height, _ = strconv.Atoi(r.FormValue("height")) } n.Data, _, _ = images.Resized(ext, n.Data, width, height) } if e := writeResponseContent(filename, mtype, bytes.NewReader(n.Data), w, r); e != nil { glog.V(2).Infoln("response write error:", e) } }