func TestRPCClientAuth(t *testing.T) {
	cl, a1, ipc := testRPCClient(t)
	defer ipc.Shutdown()
	defer cl.Close()
	defer a1.Shutdown()

	// Setup an auth key
	ipc.authKey = "foobar"

	if err := a1.Start(); err != nil {
		t.Fatalf("err: %s", err)
	}
	testutil.Yield()

	if err := cl.UserEvent("deploy", nil, false); err.Error() != authRequired {
		t.Fatalf("err: %s", err)
	}
	testutil.Yield()

	config := client.Config{Addr: ipc.listener.Addr().String(), AuthKey: "foobar"}
	rpcClient, err := client.ClientFromConfig(&config)
	if err != nil {
		t.Fatalf("err: %s", err)
	}
	defer rpcClient.Close()

	if err := rpcClient.UserEvent("deploy", nil, false); err != nil {
		t.Fatalf("err: %s", err)
	}
}
Exemple #2
0
func DoClusterStatus(c *cli.Context) {
	rpcAddress := c.GlobalString("rpc")
	rpcAuthKey := c.GlobalString("rpc-auth")
	filterNodes := parseHostArgs(c.String("hosts"))
	tagFilter, err := parseTagArgs(c.StringSlice("tag"))

	exitcode := 0
	defer func(i int) { os.Exit(i) }(exitcode)

	if err != nil {
		fmt.Println(err)
		exitcode = 1
		return
	}

	rpcConfig := client.Config{Addr: rpcAddress, AuthKey: rpcAuthKey}
	rpcClient, err := client.ClientFromConfig(&rpcConfig)
	if err != nil {
		fmt.Printf("Unable to connect to RPC at %s: %s\n", rpcAddress, err)
		exitcode = 1
		return
	}
	defer rpcClient.Close()
	members, err := rpcClient.MembersFiltered(tagFilter, "", "")
	if err != nil {
		fmt.Printf("Error retrieving members: %s\n", err)
		exitcode = 1
		return
	}

	// filter the list of members returned for only those matching our -hosts filter if present
	filteredMembers := FilterMembers(members, filterNodes)

	fmt.Printf("%d/%d hosts matching %v %v\n", len(filteredMembers), len(members), tagFilter, filterNodes)
	w := new(tabwriter.Writer)
	w.Init(os.Stdout, 0, 8, 1, '\t', 0)
	fmt.Fprintf(w, "Name\tAddr\tTags\tStatus\n")
	for _, member := range filteredMembers {
		fmt.Fprintf(w, "%s\t%s:%d\t%v\t%s\n", member.Name, member.Addr, member.Port, member.Tags, member.Status)
	}
	w.Flush()

}
Exemple #3
0
func main() {
	if DEBUG {
		log_file, _ := os.Create(DEBUG_FILE)
		defer log_file.Close()
		log.SetOutput(log_file)
		log.Println("Serf Template starting")
	}

	if len(os.Args) != 2 {
		err := errors.New("No config file")
		panic(err)
	}

	// parse directive from config file
	directives, err := ParseDirectives(os.Args[1])
	if err != nil {
		panic(err)
	}

	if DEBUG {
		log.Printf("directives: %s", directives)
	}

	// create RPC config
	rpc_config := rpc.Config{
		Addr:    directives.Rpc_addr,
		AuthKey: directives.Rpc_auth,
		Timeout: directives.Rpc_timeout,
	}
	// create connection to the RPC interface
	rpc_client, err := rpc.ClientFromConfig(&rpc_config)
	if err != nil {
		panic(err)
	}
	defer rpc_client.Close()

	// create input channel
	ch := make(chan map[string]interface{})
	_, err = rpc_client.Stream("member-join,member-failed,member-update,member-leave,member-reap", ch)
	if err != nil {
		panic(err)
	}

	for {
		// wait for signal from serf
		<-ch

		// get members' information
		members, err := rpc_client.MembersFiltered(directives.Tags, directives.Status, directives.Name)
		if err != nil {
			panic(err)
		}

		// for each templates:
		// - render template
		// - execute command if any
		for i := 0; i < len(directives.Templates); i++ {
			RenderTemplate(directives.Templates[i].Src, directives.Templates[i].Dest, members)

			if directives.Templates[i].Cmd != "" {
				cmd_args := strings.Split(directives.Templates[i].Cmd, " ")
				cmd := exec.Command(cmd_args[0], cmd_args[1:]...)
				err := cmd.Run()
				if err != nil {
					panic(err)
				}
			}
		}
	}
}
Exemple #4
0
// RPCClient returns a new Serf RPC client with the given address.
func RPCClient(addr, auth string) (*client.RPCClient, error) {
	config := client.Config{Addr: addr, AuthKey: auth}
	return client.ClientFromConfig(&config)
}
Exemple #5
0
// DoQuery will perform a query named c.Command.Name with provided parameters, and await responses.
func DoQuery(action string) func(c *cli.Context) {
	// TODO(gabe) FIXME(gabe) this is annoying. For some reason, the cli.Context.Command provided by cli is
	// empty, so we cannot extract the c.Command.Name given a context. Due to this, we need to parameterize
	// this function with the action when we setup the cli actions, so each query will know what its action
	// is. Ideally, c.Command.Name is populated to be the command name (i.e. "verify") so we can get rid of
	// this crazy closure.
	return func(c *cli.Context) {

		// allow error handlers to set exitcode and return, and still close and clients, and then call os.Exit
		// once that stuff is done.
		exitcode := 0
		timeStart := time.Now()
		defer func(i int) { os.Exit(i) }(exitcode)

		rpcAddress := c.GlobalString("rpc")
		rpcAuthKey := c.GlobalString("rpc-auth")
		timeout := c.Duration("timeout")
		if timeout == 0 {
			timeout = 10 * time.Second
		}
		args := c.Args()
		if len(args) < 1 {
			fmt.Printf("%s requires a target\n", action)
			exitcode = 1
			return
		}
		target := args[0]
		args = args[1:]

		if action == "" {
			fmt.Println("action cannot be empty!")
			exitcode = 1
			return
		}

		_, ok := config.Targets[target]
		if !ok {
			fmt.Printf("Unknown %s target %q (check your config)\n", action, target)
			exitcode = 1
			return
		}

		// arg something we allow the action script to be parameterized by. Allow it to be empty
		var arg string
		if len(args) > 0 {
			arg = args[0]
		}

		filterNodes := parseHostArgs(c.String("hosts")) // filter the query for only nodes matching these
		filterTags, err := parseTagArgs(c.StringSlice("tag"))
		if err != nil {
			fmt.Println(err)
			exitcode = 1
			return
		}
		// always restrict queries to agent=true so we only get responses from agents, not coordinating masters
		filterTags["agent"] = "true"

		//filter query for only tags matching these
		//fmt.Printf("Got filternodes %v and filtertags %v\n", filterNodes, filterTags)

		message := MessagePayload{
			Target:   target,
			Argument: arg,
		}
		messageEnc, err := message.Encode()
		if err != nil {
			fmt.Printf("Unable to encode payload: %s\n", err)
			exitcode = 1
			return
		}

		rpcConfig := client.Config{Addr: rpcAddress, AuthKey: rpcAuthKey}
		rpcClient, err := client.ClientFromConfig(&rpcConfig)
		if err != nil {
			fmt.Printf("Unable to connect to RPC at %s: %s\n", rpcAddress, err)
			exitcode = 1
			return
		}
		defer func() { rpcClient.Close() }()

		ackCh := make(chan string, CHANNEL_BUFFER)
		respCh := make(chan client.NodeResponse, CHANNEL_BUFFER)
		q := client.QueryParam{
			RequestAck: true,
			Timeout:    timeout,
			Name:       action,
			Payload:    messageEnc,
			AckCh:      ackCh,
			RespCh:     respCh,
		}
		// if we blindly set an empty map/slice here, we wont get any responses from agents :(
		// so, we only set them if there is anything of value in them
		if len(filterTags) > 0 {
			q.FilterTags = filterTags
		}
		if len(filterNodes) > 0 {
			q.FilterNodes = filterNodes
		}
		err = rpcClient.Query(&q)
		if err != nil {
			fmt.Println(err)
			exitcode = 1
			return
		}

		//TODO determine how many nodes we _should_ be hearing from, so we can display percentages...
		//TODO should this list come from collins/a source of truth? Maybe not...
		clusterMembers, err := rpcClient.MembersFiltered(filterTags, "", "")
		if err != nil {
			fmt.Println(err)
			exitcode = 1
			return
		}
		// filter by node name manually, cause MembersFiltered doesnt take a list of hosts (only a name regexp)
		expectedClusterMembers := FilterMembers(clusterMembers, filterNodes)
		fmt.Printf("%s %s %s on %d hosts matching %v %v (timeout: %s)\n", action, message.Target,
			message.Argument, len(expectedClusterMembers), filterTags, filterNodes, timeout.String())

		// track our incoming acks and responses
		acks := []string{}
		resps := []client.NodeResponse{}
		errorResponses := []QueryResponse{}

		for {
			select {
			case ack := <-ackCh:
				acks = append(acks, ack)
			case resp := <-respCh:
				queryResponse, err := DecodeQueryResponse(resp.Payload)
				if err != nil {
					fmt.Printf("Unable to decode response from %s: %s\n", resp.From, err)
					continue
				}
				status, output := "", ""
				if queryResponse.Err != nil || queryResponse.Status != 0 {
					// something actually broke in serfbort
					status = "ERROR"
					if queryResponse.Err != nil {
						output += strings.TrimSpace(*queryResponse.Err) + ": "
					}
					output += strings.TrimSpace(queryResponse.Output)
					errorResponses = append(errorResponses, queryResponse)
				} else {
					status = "OK   "
					output = strings.TrimSpace(queryResponse.Output)
				}
				fmt.Printf("%s %s: %s\n", status, resp.From, output)
				resps = append(resps, resp)
			default: //chill out, squire! no messages
			}
			if len(resps) >= len(expectedClusterMembers) {
				//fmt.Println("get outta here! we got what we came for")
				break
			}
		}
		duration := time.Since(timeStart)
		var ackPct float64 = float64(len(acks)) / float64(len(expectedClusterMembers)) * 100.0
		var respPct float64 = float64(len(resps)) / float64(len(expectedClusterMembers)) * 100.0
		var errPct float64 = float64(len(errorResponses)) / float64(len(expectedClusterMembers)) * 100.0
		//TODO this should show the actual duration of the query, not the timeout...
		fmt.Printf("%d/%d (%.1f%%) ACKed and %d/%d (%.1f%%) responded in %s\n",
			len(acks), len(expectedClusterMembers), ackPct,
			len(resps), len(expectedClusterMembers), respPct,
			duration.String())
		fmt.Printf("%d/%d (%.1f%%) reported errors\n",
			len(errorResponses), len(expectedClusterMembers), errPct)
		if len(errorResponses) > 0 {
			exitcode = 2
			return
		}

	}

}
func main() {
	config := client.Config{Addr: "127.0.0.1:7373"}
	cli, err := client.ClientFromConfig(&config)
	erro.Trata(err)
	fmt.Println(cli.Members())
}
Exemple #7
0
func connectSerfAgent(serfRPCAddress string, serfRPCAuthKey string) (*serf_client.RPCClient, error) {
	return serf_client.ClientFromConfig(&serf_client.Config{
		Addr:    serfRPCAddress,
		AuthKey: serfRPCAuthKey,
	})
}