func Serve(producer sarama.SyncProducer, topic string) { for { fmt.Print("x y: ") var x, y int fmt.Scanf("%d %d", &x, &y) m := Multiply{ X: x, Y: y, } jsonMsg, err := json.Marshal(m) if err != nil { log.Fatalln(err) } msg := sarama.ProducerMessage{ Topic: topic, Value: sarama.ByteEncoder(jsonMsg), } partition, offset, err := producer.SendMessage(&msg) if err != nil { log.Fatal(err) } else { fmt.Println("Sent msg to partition:", partition, ", offset:", offset) } } }
func produce(producer sarama.SyncProducer, c chan int, n int, s int, topic string, logger *log.Logger) { msg := &sarama.ProducerMessage{Topic: topic, Value: sarama.StringEncoder(randomString(s))} i := 0 for ; i < n; i++ { _, _, err := producer.SendMessage(msg) if err != nil { logger.Printf("FAILED to send message: %s\n", err) } } c <- i }
func (z *zipkin) pub(s *zipkincore.Span, pr sarama.SyncProducer) { t := thrift.NewTMemoryBufferLen(1024) p := thrift.NewTBinaryProtocolTransport(t) if err := s.Write(p); err != nil { return } m := &sarama.ProducerMessage{ Topic: z.opts.Topic, Value: sarama.ByteEncoder(t.Buffer.Bytes()), } pr.SendMessage(m) }
func (k *KafkaProducer) produce(producer sarama.SyncProducer) { for { metric := <-k.metricsChannel json, err := json.MarshalIndent(metric, "", " ") if err != nil { return } msg := &sarama.ProducerMessage{Topic: "loadbalancer.all", Value: sarama.StringEncoder(json)} _, _, err = producer.SendMessage(msg) if err != nil { k.Log.Error("error sending to Kafka ") } } }
// Publish message func sendMessage(producer sarama.SyncProducer, topic string, payload string) { message := sarama.ProducerMessage{ // The Kafka topic for this message Topic: topic, // The actual message to store in Kafka Value: sarama.StringEncoder(payload), // No message key, so messages will be distributed randomly over partitions } // Send Message partition, offset, err := producer.SendMessage(&message) if err != nil { logger.Printf("Error sending data: %s\n", err) } else { logger.Printf("[%s/%d/%d] Message successfully published\n", topic, partition, offset) } }
// Note that unfortunately this does not actually produce in batches yet. We // should theoretically be able to with Kafka, but the sarama interface for a // `SyncProducer` currently seems overly limited. func processBatch(producer sarama.SyncProducer, topic string, events []*stripe.Event) error { for _, event := range events { data, err := json.Marshal(event) if err != nil { return err } id := event.Data.Obj["id"] if id == nil { log.Printf("Found event with nil data ID, type is %v", event.Type) } // TODO: Verify that Kafka does indeed perform log compaction per // partition key (as opposed to some other type of "key"). The docs // aren't exactly clear on this point. key := "" if id != nil { key = id.(string) } message := &sarama.ProducerMessage{ Topic: topic, Key: sarama.StringEncoder(key), Value: sarama.ByteEncoder(data), } //start := time.Now() //partition, offset, err := producer.SendMessage(message) _, _, err = producer.SendMessage(message) if err != nil { return err } else { //log.Printf("> Message sent to partition %d at offset %d in %v\n", //partition, offset, time.Now().Sub(start)) } } return nil }
func PublishSync(input chan *FileEvent, source string, isRetryer bool) { log.Debug("publishSync loop") clientConfig := sarama.NewConfig() clientConfig.Producer.RequiredAcks = sarama.WaitForAll clientConfig.Producer.Compression = sarama.CompressionSnappy clientConfig.Producer.Partitioner = sarama.NewRoundRobinPartitioner clientConfig.Producer.Retry.Max = 10 topic := kafkaTopic key := hashKey if isRetryer { topic = retryTopic } //brokerList := []string{"127.0.0.1:9092"} var producer sarama.SyncProducer var err error for { producer, err = sarama.NewSyncProducer(brokerList, clientConfig) if err != nil { log.Error("Sync: Failed to start Sarama producer: ", err) log.Info("waiting...") time.Sleep(1 * time.Second) } else { break } } defer func() { if err := producer.Close(); err != nil { log.Error("Failed to shutdown producer cleanly", err) } }() // if retryer, use retryer backup, others use Registrar var recorder Recorder if isRetryer { // set to global retryer retryRecorder := &RetryRecorder{file: mainRetryer.vernier} recorder = retryRecorder defer retryRecorder.file.Close() } else { registrar := &Registrar{source: source, dir: REGISTRAR_DIR} if _, err := registrar.OpenRecord(registrar.dir); err != nil { log.Error("PublishSync open record failed, error:", err) os.Exit(2) } recorder = registrar } genMessage := func(rawMessage string) string { return rawMessage } // retryer message sample: 0 this is a sample message // 0 means, haven't retried succeed // 1 means have been sended if isRetryer { genMessage = func(rawMessage string) string { // 0|1 raw_name_of_log_file log_msg rawMessage = rawMessage[2:] idx := strings.Index(rawMessage, " ") return rawMessage[idx+1:] } } for event := range input { log.Debugf("%v, %v, %v, %v\n", *event.Source, *event.Text, event.Line, event.Offset) // if failed, retry send messge until succeed key = strconv.FormatInt(event.Offset, 10) rawMessage := *event.Text if isRetryer { if retryTopic != kafkaTopic { topic = retryTopic } else { baseName := getSourceName(rawMessage) if len(topicmap) > 0 { tmpTopic := genTopic(baseName, topicmap) if tmpTopic != "" { topic = tmpTopic } } } } message := genMessage(*event.Text) if rawMessage[0] == '1' { log.Debug("message[%s] have been seeded", rawMessage) continue } for { partition, offset, err := producer.SendMessage(&sarama.ProducerMessage{ Topic: topic, Key: sarama.StringEncoder(key), Value: sarama.StringEncoder(message), }) if err != nil { log.Errorf("Failed: %s, %d, %d\n", *event.Source, event.Line, event.Offset) time.Sleep(3 * time.Second) } else { log.Debugf("OK: %d, %d, %s\n", partition, offset, *event.Source) recorder.RecordSucceed(event.Offset, event.RawBytes) break } } } }
func startIndexDumper(producer sarama.SyncProducer, wg *sync.WaitGroup) chan<- *indexDumperMessage { ch := make(chan *indexDumperMessage, 300) // 5 min wg.Add(1) go func() { defer wg.Done() buf8 := make([]byte, 8) builder := flatbuffers.NewBuilder(1024 * 1024) for { msg := <-ch if msg == nil { log.Println("exiting index dumper") return } start := time.Now() t, index := msg.t, msg.index // log.Printf("index dumper got index for %d", t.Unix()) var tags []string for tag := range index { tags = append(tags, tag) } sort.Strings(tags) builder.Reset() var fbtags []flatbuffers.UOffsetT for _, tag := range tags { name := builder.CreateString(tag) cnt := 0 builder.StartVector(flatbuffers.SizeInt64, 0, 0) for partition, d := range index[tag] { for di := d.head; di != nil; di = di.next { for i := 0; i < di.cnt; i++ { builder.PrependInt64(encodePartitionAndOffset(partition, di.vals[i])) cnt++ } } } offsetsVector := builder.EndVector(cnt) fb.TagStart(builder) fb.TagAddName(builder, name) fb.TagAddOffsets(builder, offsetsVector) fbtags = append(fbtags, fb.TagEnd(builder)) } fb.IndexStartTagsVector(builder, len(fbtags)) for _, offset := range fbtags { builder.PrependUOffsetT(offset) } tagsVector := builder.EndVector(len(fbtags)) fb.IndexStart(builder) fb.IndexAddTags(builder, tagsVector) builder.Finish(fb.IndexEnd(builder)) encoded := builder.FinishedBytes() binary.LittleEndian.PutUint64(buf8, uint64(t.Unix())) statIncrementTook(&stat.idxSerializeTook, start) statIncrementSize(&stat.idxSendToKafkaSize, len(encoded)) start = time.Now() _, _, err := producer.SendMessage(&sarama.ProducerMessage{ Topic: indexTopic, Key: sarama.ByteEncoder(buf8), Value: sarama.ByteEncoder(encoded), }) statIncrementTook(&stat.idxSendToKafkaTook, start) if err != nil { log.Printf("failed to store message: %v", err) } // log.Printf("finished serializing index for %d, %d msgs, %d tags, %d offsets", // t.Unix(), msg.msgs, len(tags), msg.offsets) } }() return ch }