aboutsummaryrefslogtreecommitdiff
path: root/weed/shell/command_cluster_check.go
diff options
context:
space:
mode:
authoryulai.li <blacktear23@gmail.com>2022-06-26 22:43:37 +0800
committeryulai.li <blacktear23@gmail.com>2022-06-26 22:43:37 +0800
commit46e0b629e529f3aff535f90dd25eb719adf1c0d0 (patch)
tree734125b48b6d96f8796a2b89b924312cd169ef0e /weed/shell/command_cluster_check.go
parenta5bd0b3a1644a77dcc0b9ff41c4ce8eb3ea0d566 (diff)
parentdc59ccd110a321db7d0b0480631aa95a3d9ba7e6 (diff)
downloadseaweedfs-46e0b629e529f3aff535f90dd25eb719adf1c0d0.tar.xz
seaweedfs-46e0b629e529f3aff535f90dd25eb719adf1c0d0.zip
Update tikv client version and add one PC support
Diffstat (limited to 'weed/shell/command_cluster_check.go')
-rw-r--r--weed/shell/command_cluster_check.go249
1 files changed, 249 insertions, 0 deletions
diff --git a/weed/shell/command_cluster_check.go b/weed/shell/command_cluster_check.go
new file mode 100644
index 000000000..e363239e7
--- /dev/null
+++ b/weed/shell/command_cluster_check.go
@@ -0,0 +1,249 @@
+package shell
+
+import (
+ "context"
+ "flag"
+ "fmt"
+ "github.com/chrislusf/seaweedfs/weed/cluster"
+ "github.com/chrislusf/seaweedfs/weed/pb"
+ "github.com/chrislusf/seaweedfs/weed/pb/filer_pb"
+ "github.com/chrislusf/seaweedfs/weed/pb/master_pb"
+ "github.com/chrislusf/seaweedfs/weed/pb/volume_server_pb"
+ "io"
+)
+
+func init() {
+ Commands = append(Commands, &commandClusterCheck{})
+}
+
+type commandClusterCheck struct {
+}
+
+func (c *commandClusterCheck) Name() string {
+ return "cluster.check"
+}
+
+func (c *commandClusterCheck) Help() string {
+ return `check current cluster network connectivity
+
+ cluster.check
+
+`
+}
+
+func (c *commandClusterCheck) Do(args []string, commandEnv *CommandEnv, writer io.Writer) (err error) {
+
+ clusterPsCommand := flag.NewFlagSet(c.Name(), flag.ContinueOnError)
+ if err = clusterPsCommand.Parse(args); err != nil {
+ return nil
+ }
+
+ // collect topology information
+ topologyInfo, volumeSizeLimitMb, err := collectTopologyInfo(commandEnv, 0)
+ if err != nil {
+ return err
+ }
+ fmt.Fprintf(writer, "Topology volumeSizeLimit:%d MB%s\n", volumeSizeLimitMb, diskInfosToString(topologyInfo.DiskInfos))
+
+ emptyDiskTypeDiskInfo, emptyDiskTypeFound := topologyInfo.DiskInfos[""]
+ hddDiskTypeDiskInfo, hddDiskTypeFound := topologyInfo.DiskInfos["hdd"]
+ if !emptyDiskTypeFound && !hddDiskTypeFound {
+ return fmt.Errorf("Need to a hdd disk type!")
+ }
+ if emptyDiskTypeFound && emptyDiskTypeDiskInfo.MaxVolumeCount == 0 || hddDiskTypeFound && hddDiskTypeDiskInfo.MaxVolumeCount == 0 {
+ return fmt.Errorf("Need to a hdd disk type!")
+ }
+
+ // collect filers
+ var filers []pb.ServerAddress
+ err = commandEnv.MasterClient.WithClient(false, func(client master_pb.SeaweedClient) error {
+ resp, err := client.ListClusterNodes(context.Background(), &master_pb.ListClusterNodesRequest{
+ ClientType: cluster.FilerType,
+ FilerGroup: *commandEnv.option.FilerGroup,
+ })
+
+ for _, node := range resp.ClusterNodes {
+ filers = append(filers, pb.ServerAddress(node.Address))
+ }
+ return err
+ })
+ if err != nil {
+ return
+ }
+ fmt.Fprintf(writer, "the cluster has %d filers: %+v\n", len(filers), filers)
+
+ // collect volume servers
+ var volumeServers []pb.ServerAddress
+ t, _, err := collectTopologyInfo(commandEnv, 0)
+ if err != nil {
+ return err
+ }
+ for _, dc := range t.DataCenterInfos {
+ for _, r := range dc.RackInfos {
+ for _, dn := range r.DataNodeInfos {
+ volumeServers = append(volumeServers, pb.NewServerAddressFromDataNode(dn))
+ }
+ }
+ }
+ fmt.Fprintf(writer, "the cluster has %d volume servers: %+v\n", len(volumeServers), volumeServers)
+
+ // collect all masters
+ var masters []pb.ServerAddress
+ for _, master := range commandEnv.MasterClient.GetMasters() {
+ masters = append(masters, master)
+ }
+
+ // check from master to volume servers
+ for _, master := range masters {
+ for _, volumeServer := range volumeServers {
+ fmt.Fprintf(writer, "checking master %s to volume server %s ... ", string(master), string(volumeServer))
+ err := pb.WithMasterClient(false, master, commandEnv.option.GrpcDialOption, func(client master_pb.SeaweedClient) error {
+ pong, err := client.Ping(context.Background(), &master_pb.PingRequest{
+ Target: string(volumeServer),
+ TargetType: cluster.VolumeServerType,
+ })
+ if err == nil {
+ printTiming(writer, pong.StartTimeNs, pong.RemoteTimeNs, pong.StopTimeNs)
+ }
+ return err
+ })
+ if err != nil {
+ fmt.Fprintf(writer, "%v\n", err)
+ }
+ }
+ }
+
+ // check between masters
+ for _, sourceMaster := range masters {
+ for _, targetMaster := range masters {
+ if sourceMaster == targetMaster {
+ continue
+ }
+ fmt.Fprintf(writer, "checking master %s to %s ... ", string(sourceMaster), string(targetMaster))
+ err := pb.WithMasterClient(false, sourceMaster, commandEnv.option.GrpcDialOption, func(client master_pb.SeaweedClient) error {
+ pong, err := client.Ping(context.Background(), &master_pb.PingRequest{
+ Target: string(targetMaster),
+ TargetType: cluster.MasterType,
+ })
+ if err == nil {
+ printTiming(writer, pong.StartTimeNs, pong.RemoteTimeNs, pong.StopTimeNs)
+ }
+ return err
+ })
+ if err != nil {
+ fmt.Fprintf(writer, "%v\n", err)
+ }
+ }
+ }
+
+ // check from volume servers to masters
+ for _, volumeServer := range volumeServers {
+ for _, master := range masters {
+ fmt.Fprintf(writer, "checking volume server %s to master %s ... ", string(volumeServer), string(master))
+ err := pb.WithVolumeServerClient(false, volumeServer, commandEnv.option.GrpcDialOption, func(client volume_server_pb.VolumeServerClient) error {
+ pong, err := client.Ping(context.Background(), &volume_server_pb.PingRequest{
+ Target: string(master),
+ TargetType: cluster.MasterType,
+ })
+ if err == nil {
+ printTiming(writer, pong.StartTimeNs, pong.RemoteTimeNs, pong.StopTimeNs)
+ }
+ return err
+ })
+ if err != nil {
+ fmt.Fprintf(writer, "%v\n", err)
+ }
+ }
+ }
+
+ // check from filers to masters
+ for _, filer := range filers {
+ for _, master := range masters {
+ fmt.Fprintf(writer, "checking filer %s to master %s ... ", string(filer), string(master))
+ err := pb.WithFilerClient(false, filer, commandEnv.option.GrpcDialOption, func(client filer_pb.SeaweedFilerClient) error {
+ pong, err := client.Ping(context.Background(), &filer_pb.PingRequest{
+ Target: string(master),
+ TargetType: cluster.MasterType,
+ })
+ if err == nil {
+ printTiming(writer, pong.StartTimeNs, pong.RemoteTimeNs, pong.StopTimeNs)
+ }
+ return err
+ })
+ if err != nil {
+ fmt.Fprintf(writer, "%v\n", err)
+ }
+ }
+ }
+
+ // check from filers to volume servers
+ for _, filer := range filers {
+ for _, volumeServer := range volumeServers {
+ fmt.Fprintf(writer, "checking filer %s to volume server %s ... ", string(filer), string(volumeServer))
+ err := pb.WithFilerClient(false, filer, commandEnv.option.GrpcDialOption, func(client filer_pb.SeaweedFilerClient) error {
+ pong, err := client.Ping(context.Background(), &filer_pb.PingRequest{
+ Target: string(volumeServer),
+ TargetType: cluster.VolumeServerType,
+ })
+ if err == nil {
+ printTiming(writer, pong.StartTimeNs, pong.RemoteTimeNs, pong.StopTimeNs)
+ }
+ return err
+ })
+ if err != nil {
+ fmt.Fprintf(writer, "%v\n", err)
+ }
+ }
+ }
+
+ // check between volume servers
+ for _, sourceVolumeServer := range volumeServers {
+ for _, targetVolumeServer := range volumeServers {
+ if sourceVolumeServer == targetVolumeServer {
+ continue
+ }
+ fmt.Fprintf(writer, "checking volume server %s to %s ... ", string(sourceVolumeServer), string(targetVolumeServer))
+ err := pb.WithVolumeServerClient(false, sourceVolumeServer, commandEnv.option.GrpcDialOption, func(client volume_server_pb.VolumeServerClient) error {
+ pong, err := client.Ping(context.Background(), &volume_server_pb.PingRequest{
+ Target: string(targetVolumeServer),
+ TargetType: cluster.VolumeServerType,
+ })
+ if err == nil {
+ printTiming(writer, pong.StartTimeNs, pong.RemoteTimeNs, pong.StopTimeNs)
+ }
+ return err
+ })
+ if err != nil {
+ fmt.Fprintf(writer, "%v\n", err)
+ }
+ }
+ }
+
+ // check between filers, and need to connect to itself
+ for _, sourceFiler := range filers {
+ for _, targetFiler := range filers {
+ fmt.Fprintf(writer, "checking filer %s to %s ... ", string(sourceFiler), string(targetFiler))
+ err := pb.WithFilerClient(false, sourceFiler, commandEnv.option.GrpcDialOption, func(client filer_pb.SeaweedFilerClient) error {
+ pong, err := client.Ping(context.Background(), &filer_pb.PingRequest{
+ Target: string(targetFiler),
+ TargetType: cluster.FilerType,
+ })
+ if err == nil {
+ printTiming(writer, pong.StartTimeNs, pong.RemoteTimeNs, pong.StopTimeNs)
+ }
+ return err
+ })
+ if err != nil {
+ fmt.Fprintf(writer, "%v\n", err)
+ }
+ }
+ }
+
+ return nil
+}
+
+func printTiming(writer io.Writer, startNs, remoteNs, stopNs int64) {
+ roundTripTimeMs := float32(stopNs-startNs) / 1000000
+ deltaTimeMs := float32(remoteNs-(startNs+stopNs)/2) / 1000000
+ fmt.Fprintf(writer, "ok round trip %.3fms clock delta %.3fms\n", roundTripTimeMs, deltaTimeMs)
+}