// processBatches continually drains the given batcher and writes the batches to the database. func (s *Service) processBatches(batcher *tsdb.PointBatcher) { defer s.wg.Done() for { select { case batch := <-batcher.Out(): if err := s.PointsWriter.WritePoints(s.database, s.retentionPolicy, models.ConsistencyLevelAny, batch); err == nil { s.statMap.Add(statBatchesTransmitted, 1) s.statMap.Add(statPointsTransmitted, int64(len(batch))) } else { s.logger.Printf("failed to write point batch to database %q: %s", s.database, err) s.statMap.Add(statBatchesTransmitFail, 1) } case <-s.done: return } } }
// processBatches continually drains the given batcher and writes the batches to the database. func (s *Service) processBatches(batcher *tsdb.PointBatcher) { for { select { case <-s.done: return case batch := <-batcher.Out(): // Will attempt to create database if not yet created. if err := s.createInternalStorage(); err != nil { s.Logger.Info(fmt.Sprintf("Required database %s does not yet exist: %s", s.Database, err.Error())) continue } if err := s.PointsWriter.WritePoints(s.Database, s.RetentionPolicy, models.ConsistencyLevelAny, batch); err == nil { atomic.AddInt64(&s.stats.BatchesTransmitted, 1) atomic.AddInt64(&s.stats.PointsTransmitted, int64(len(batch))) } else { s.Logger.Info(fmt.Sprintf("failed to write point batch to database %q: %s", s.Database, err)) atomic.AddInt64(&s.stats.BatchesTransmitFail, 1) } } } }