// GetPlots retrieves time series data from origin based on a query and a time interval. func (c *RRDConnector) GetPlots(query *plot.Query) ([]*plot.Series, error) { var ( results []*plot.Series xport *rrd.Exporter ) if len(query.Series) == 0 { return nil, fmt.Errorf("rrd[%s]: requested series list is empty", c.name) } graph := rrd.NewGrapher() if c.daemon != "" { graph.SetDaemon(c.daemon) } xport = rrd.NewExporter() if c.daemon != "" { xport.SetDaemon(c.daemon) } step := time.Duration(0) for _, s := range query.Series { filePath := strings.Replace(c.metrics[s.Source][s.Metric].FilePath, ":", "\\:", -1) graph.Def(s.Name+"-def0", filePath, c.metrics[s.Source][s.Metric].Dataset, c.metrics[s.Source][s.Metric].Cf) graph.CDef(s.Name, s.Name+"-def0") // Set plots request xport.Def(s.Name+"-def0", filePath, c.metrics[s.Source][s.Metric].Dataset, c.metrics[s.Source][s.Metric].Cf) xport.CDef(s.Name, s.Name+"-def0") xport.XportDef(s.Name, s.Name) if c.metrics[s.Source][s.Metric].Step > step { step = c.metrics[s.Source][s.Metric].Step } } // Get plots if step == 0 { step = query.EndTime.Sub(query.StartTime) / time.Duration(config.DefaultPlotSample) } data := rrd.XportResult{} data, err := xport.Xport(query.StartTime, query.EndTime, step) if err != nil { return nil, err } for idx, name := range data.Legends { series := &plot.Series{ Name: name, } // FIXME: skip last garbage entry (see https://github.com/ziutek/rrd/pull/13) for i := 0; i < data.RowCnt-1; i++ { series.Plots = append(series.Plots, plot.Plot{ Time: query.StartTime.Add(data.Step * time.Duration(i)), Value: plot.Value(data.ValueAt(idx, i)), }) } results = append(results, series) } data.FreeValues() return results, nil }
func main() { // Create const ( dbfile = "/tmp/test.rrd" step = 1 heartbeat = 2 * step ) c := rrd.NewCreator(dbfile, time.Now(), step) c.RRA("AVERAGE", 0.5, 1, 100) c.RRA("AVERAGE", 0.5, 5, 100) c.DS("cnt", "COUNTER", heartbeat, 0, 100) c.DS("g", "GAUGE", heartbeat, 0, 60) err := c.Create(true) if err != nil { panic(err) } // Update u := rrd.NewUpdater(dbfile) for i := 0; i < 10; i++ { time.Sleep(step * time.Second) err := u.Update(time.Now(), i, 1.5*float64(i)) if err != nil { panic(err) } } // Update with cache for i := 10; i < 20; i++ { time.Sleep(step * time.Second) u.Cache(time.Now(), i, 2*float64(i)) } err = u.Update() if err != nil { panic(err) } // Info inf, err := rrd.Info(dbfile) if err != nil { panic(err) } for k, v := range inf { fmt.Printf("%s (%T): %v\n", k, v, v) } // Graph g := rrd.NewGrapher() g.SetTitle("Test") g.SetVLabel("some variable") g.SetSize(800, 300) g.SetWatermark("some watermark") g.Def("v1", dbfile, "g", "AVERAGE") g.Def("v2", dbfile, "cnt", "AVERAGE") g.VDef("max1", "v1,MAXIMUM") g.VDef("avg2", "v2,AVERAGE") g.Line(1, "v1", "ff0000", "var 1") g.Area("v2", "0000ff", "var 2") g.GPrintT("max1", "max1 at %c") g.GPrint("avg2", "avg2=%lf") g.PrintT("max1", "max1 at %c") g.Print("avg2", "avg2=%lf") now := time.Now() i, err := g.SaveGraph("/tmp/test_rrd1.png", now.Add(-20*time.Second), now) fmt.Printf("%+v\n", i) if err != nil { panic(err) } i, buf, err := g.Graph(now.Add(-20*time.Second), now) fmt.Printf("%+v\n", i) if err != nil { panic(err) } err = ioutil.WriteFile("/tmp/test_rrd2.png", buf, 0666) if err != nil { panic(err) } // Fetch end := time.Unix(int64(inf["last_update"].(uint)), 0) start := end.Add(-20 * step * time.Second) fmt.Printf("Fetch Params:\n") fmt.Printf("Start: %s\n", start) fmt.Printf("End: %s\n", end) fmt.Printf("Step: %s\n", step*time.Second) fetchRes, err := rrd.Fetch(dbfile, "AVERAGE", start, end, step*time.Second) if err != nil { panic(err) } defer fetchRes.FreeValues() fmt.Printf("FetchResult:\n") fmt.Printf("Start: %s\n", fetchRes.Start) fmt.Printf("End: %s\n", fetchRes.End) fmt.Printf("Step: %s\n", fetchRes.Step) for _, dsName := range fetchRes.DsNames { fmt.Printf("\t%s", dsName) } fmt.Printf("\n") row := 0 for ti := fetchRes.Start.Add(fetchRes.Step); ti.Before(end) || ti.Equal(end); ti = ti.Add(fetchRes.Step) { fmt.Printf("%s / %d", ti, ti.Unix()) for i := 0; i < len(fetchRes.DsNames); i++ { v := fetchRes.ValueAt(i, row) fmt.Printf("\t%e", v) } fmt.Printf("\n") row++ } // Xport end = time.Unix(int64(inf["last_update"].(uint)), 0) start = end.Add(-20 * step * time.Second) fmt.Printf("Xport Params:\n") fmt.Printf("Start: %s\n", start) fmt.Printf("End: %s\n", end) fmt.Printf("Step: %s\n", step*time.Second) e := rrd.NewExporter() e.Def("def1", dbfile, "cnt", "AVERAGE") e.Def("def2", dbfile, "g", "AVERAGE") e.CDef("vdef1", "def1,def2,+") e.XportDef("def1", "cnt") e.XportDef("def2", "g") e.XportDef("vdef1", "sum") xportRes, err := e.Xport(start, end, step*time.Second) if err != nil { panic(err) } defer xportRes.FreeValues() fmt.Printf("XportResult:\n") fmt.Printf("Start: %s\n", xportRes.Start) fmt.Printf("End: %s\n", xportRes.End) fmt.Printf("Step: %s\n", xportRes.Step) for _, legend := range xportRes.Legends { fmt.Printf("\t%s", legend) } fmt.Printf("\n") row = 0 for ti := xportRes.Start.Add(xportRes.Step); ti.Before(end) || ti.Equal(end); ti = ti.Add(xportRes.Step) { fmt.Printf("%s / %d", ti, ti.Unix()) for i := 0; i < len(xportRes.Legends); i++ { v := xportRes.ValueAt(i, row) fmt.Printf("\t%e", v) } fmt.Printf("\n") row++ } }