func (f *NullFS) Read(r *srv.Req) { var count int ninep.InitRread(r.Rc, r.Tc.Count) fid := r.Fid.Aux.(*Fid) if fid.Qid.Path == Qroot { var dirents []byte for path, v := range dirQids { d := &ninep.Dir{ Qid: *v, Type: uint16(v.Type), Mode: uint32(v.Type) | v.Version, Name: path, Uid: "root", Gid: "root", } b := ninep.PackDir(d, true) dirents = append(dirents, b...) count += len(b) } // TODO: put this boilerplate into a helper function. switch { case r.Tc.Offset > uint64(len(dirents)): count = 0 case len(dirents[r.Tc.Offset:]) > int(r.Tc.Count): count = int(r.Tc.Count) default: count = len(dirents[r.Tc.Offset:]) } if count == 0 && int(r.Tc.Offset) < len(dirents) && len(dirents) > 0 { r.RespondError(&ninep.Error{Err: "too small read size for dir entry", Errornum: ninep.EINVAL}) return } copy(r.Rc.Data, dirents[r.Tc.Offset:int(r.Tc.Offset)+count]) } else { if fid.Qid.Path == Qzero { count = int(r.Tc.Count) } } ninep.SetRreadCount(r.Rc, uint32(count)) r.Respond() }
// Write handles writes for writeable files and always succeeds. // Only the null files has w so we don't bother checking Path. func (f *NullFS) Write(r *srv.Req) { var count uint32 ninep.SetRreadCount(r.Rc, uint32(count)) r.Respond() }
func (u *Ufs) Read(req *srv.Req) { dbg := u.Debuglevel&srv.DbgLogFcalls != 0 fid := req.Fid.Aux.(*Fid) tc := req.Tc rc := req.Rc err := fid.stat() if err != nil { req.RespondError(err) return } ninep.InitRread(rc, tc.Count) var count int var e error if fid.st.IsDir() { if tc.Offset == 0 { var e error // If we got here, it was open. Can't really seek // in most cases, just close and reopen it. fid.file.Close() if fid.file, e = os.OpenFile(fid.path, omode2uflags(req.Fid.Omode), 0); e != nil { req.RespondError(toError(e)) return } if fid.dirs, e = fid.file.Readdir(-1); e != nil { req.RespondError(toError(e)) return } if dbg { log.Printf("Read: read %d entries", len(fid.dirs)) } fid.dirents = nil fid.direntends = nil for i := 0; i < len(fid.dirs); i++ { path := fid.path + "/" + fid.dirs[i].Name() st, err := dir2Dir(path, fid.dirs[i], req.Conn.Dotu, req.Conn.Srv.Upool) if err != nil { if dbg { log.Printf("dbg: stat of %v: %v", path, err) } continue } if dbg { log.Printf("Stat: %v is %v", path, st) } b := ninep.PackDir(st, req.Conn.Dotu) fid.dirents = append(fid.dirents, b...) count += len(b) fid.direntends = append(fid.direntends, count) if dbg { log.Printf("fid.direntends is %v\n", fid.direntends) } } } switch { case tc.Offset > uint64(len(fid.dirents)): count = 0 case len(fid.dirents[tc.Offset:]) > int(tc.Count): count = int(tc.Count) default: count = len(fid.dirents[tc.Offset:]) } if dbg { log.Printf("readdir: count %v @ offset %v", count, tc.Offset) } nextend := sort.SearchInts(fid.direntends, int(tc.Offset)+count) if nextend < len(fid.direntends) { if fid.direntends[nextend] > int(tc.Offset)+count { if nextend > 0 { count = fid.direntends[nextend-1] - int(tc.Offset) } else { count = 0 } } } if dbg { log.Printf("readdir: count adjusted %v @ offset %v", count, tc.Offset) } if count == 0 && int(tc.Offset) < len(fid.dirents) && len(fid.dirents) > 0 { req.RespondError(&ninep.Error{"too small read size for dir entry", ninep.EINVAL}) return } copy(rc.Data, fid.dirents[tc.Offset:int(tc.Offset)+count]) } else { count, e = fid.file.ReadAt(rc.Data, int64(tc.Offset)) if e != nil && e != io.EOF { req.RespondError(toError(e)) return } } ninep.SetRreadCount(rc, uint32(count)) req.Respond() }