diff options
Diffstat (limited to 'go/topology')
| -rw-r--r-- | go/topology/node.go | 6 | ||||
| -rw-r--r-- | go/topology/node_list.go | 6 | ||||
| -rw-r--r-- | go/topology/topology.go | 5 | ||||
| -rw-r--r-- | go/topology/topology_compact.go | 22 | ||||
| -rw-r--r-- | go/topology/topology_event_handling.go | 8 | ||||
| -rw-r--r-- | go/topology/volume_layout.go | 10 |
6 files changed, 28 insertions, 29 deletions
diff --git a/go/topology/node.go b/go/topology/node.go index d61f01244..b2414032b 100644 --- a/go/topology/node.go +++ b/go/topology/node.go @@ -2,7 +2,7 @@ package topology import ( "code.google.com/p/weed-fs/go/storage" - "fmt" + "log" ) type NodeId string @@ -155,7 +155,7 @@ func (n *NodeImpl) LinkChildNode(node Node) { n.UpAdjustVolumeCountDelta(node.GetVolumeCount()) n.UpAdjustActiveVolumeCountDelta(node.GetActiveVolumeCount()) node.SetParent(n) - fmt.Println(n, "adds child", node.Id()) + log.Println(n, "adds child", node.Id()) } } @@ -167,7 +167,7 @@ func (n *NodeImpl) UnlinkChildNode(nodeId NodeId) { n.UpAdjustVolumeCountDelta(-node.GetVolumeCount()) n.UpAdjustActiveVolumeCountDelta(-node.GetActiveVolumeCount()) n.UpAdjustMaxVolumeCountDelta(-node.GetMaxVolumeCount()) - fmt.Println(n, "removes", node, "volumeCount =", n.activeVolumeCount) + log.Println(n, "removes", node, "volumeCount =", n.activeVolumeCount) } } diff --git a/go/topology/node_list.go b/go/topology/node_list.go index 2be90b123..9dfc5fded 100644 --- a/go/topology/node_list.go +++ b/go/topology/node_list.go @@ -2,7 +2,7 @@ package topology import ( "code.google.com/p/weed-fs/go/storage" - "fmt" + "log" "math/rand" ) @@ -58,7 +58,7 @@ func (nl *NodeList) RandomlyPickN(count int, minSpace int, firstNodeName string) list[r], list[i-1] = list[i-1], list[r] } if firstNodeName != "" { - list[0] = *preferredNode + list[0] = *preferredNode } return list[:count], true } @@ -70,7 +70,7 @@ func (nl *NodeList) ReserveOneVolume(randomVolumeIndex int, vid storage.VolumeId randomVolumeIndex -= freeSpace } else { if node.IsDataNode() && node.FreeSpace() > 0 { - fmt.Println("vid =", vid, " assigned to node =", node, ", freeSpace =", node.FreeSpace()) + log.Println("vid =", vid, " assigned to node =", node, ", freeSpace =", node.FreeSpace()) return true, node.(*DataNode) } children := node.Children() diff --git a/go/topology/topology.go b/go/topology/topology.go index e488319d1..b3223c5b4 100644 --- a/go/topology/topology.go +++ b/go/topology/topology.go @@ -4,7 +4,6 @@ import ( "code.google.com/p/weed-fs/go/sequence" "code.google.com/p/weed-fs/go/storage" "errors" - "fmt" "io/ioutil" "log" "math/rand" @@ -74,7 +73,7 @@ func (t *Topology) Lookup(vid storage.VolumeId) []*DataNode { func (t *Topology) RandomlyReserveOneVolume(dataCenter string) (bool, *DataNode, *storage.VolumeId) { if t.FreeSpace() <= 0 { - fmt.Println("Topology does not have free space left!") + log.Println("Topology does not have free space left!") return false, nil, nil } vid := t.NextVolumeId() @@ -103,7 +102,7 @@ func (t *Topology) PickForWrite(repType storage.ReplicationType, count int, data func (t *Topology) GetVolumeLayout(repType storage.ReplicationType) *VolumeLayout { replicationTypeIndex := repType.GetReplicationLevelIndex() if t.replicaType2VolumeLayout[replicationTypeIndex] == nil { - fmt.Println("adding replication type", repType) + log.Println("adding replication type", repType) t.replicaType2VolumeLayout[replicationTypeIndex] = NewVolumeLayout(repType, t.volumeSizeLimit, t.pulse) } return t.replicaType2VolumeLayout[replicationTypeIndex] diff --git a/go/topology/topology_compact.go b/go/topology/topology_compact.go index 7215edc4e..1e5c4dd11 100644 --- a/go/topology/topology_compact.go +++ b/go/topology/topology_compact.go @@ -5,7 +5,7 @@ import ( "code.google.com/p/weed-fs/go/util" "encoding/json" "errors" - "fmt" + "log" "net/url" "time" ) @@ -14,12 +14,12 @@ func batchVacuumVolumeCheck(vl *VolumeLayout, vid storage.VolumeId, locationlist ch := make(chan bool, locationlist.Length()) for index, dn := range locationlist.list { go func(index int, url string, vid storage.VolumeId) { - //fmt.Println(index, "Check vacuuming", vid, "on", dn.Url()) + //log.Println(index, "Check vacuuming", vid, "on", dn.Url()) if e, ret := vacuumVolume_Check(url, vid, garbageThreshold); e != nil { - //fmt.Println(index, "Error when checking vacuuming", vid, "on", url, e) + //log.Println(index, "Error when checking vacuuming", vid, "on", url, e) ch <- false } else { - //fmt.Println(index, "Checked vacuuming", vid, "on", url, "needVacuum", ret) + //log.Println(index, "Checked vacuuming", vid, "on", url, "needVacuum", ret) ch <- ret } }(index, dn.Url(), vid) @@ -41,12 +41,12 @@ func batchVacuumVolumeCompact(vl *VolumeLayout, vid storage.VolumeId, locationli ch := make(chan bool, locationlist.Length()) for index, dn := range locationlist.list { go func(index int, url string, vid storage.VolumeId) { - fmt.Println(index, "Start vacuuming", vid, "on", url) + log.Println(index, "Start vacuuming", vid, "on", url) if e := vacuumVolume_Compact(url, vid); e != nil { - fmt.Println(index, "Error when vacuuming", vid, "on", url, e) + log.Println(index, "Error when vacuuming", vid, "on", url, e) ch <- false } else { - fmt.Println(index, "Complete vacuuming", vid, "on", url) + log.Println(index, "Complete vacuuming", vid, "on", url) ch <- true } }(index, dn.Url(), vid) @@ -65,12 +65,12 @@ func batchVacuumVolumeCompact(vl *VolumeLayout, vid storage.VolumeId, locationli func batchVacuumVolumeCommit(vl *VolumeLayout, vid storage.VolumeId, locationlist *VolumeLocationList) bool { isCommitSuccess := true for _, dn := range locationlist.list { - fmt.Println("Start Commiting vacuum", vid, "on", dn.Url()) + log.Println("Start Commiting vacuum", vid, "on", dn.Url()) if e := vacuumVolume_Commit(dn.Url(), vid); e != nil { - fmt.Println("Error when committing vacuum", vid, "on", dn.Url(), e) + log.Println("Error when committing vacuum", vid, "on", dn.Url(), e) isCommitSuccess = false } else { - fmt.Println("Complete Commiting vacuum", vid, "on", dn.Url()) + log.Println("Complete Commiting vacuum", vid, "on", dn.Url()) } } if isCommitSuccess { @@ -104,7 +104,7 @@ func vacuumVolume_Check(urlLocation string, vid storage.VolumeId, garbageThresho values.Add("garbageThreshold", garbageThreshold) jsonBlob, err := util.Post("http://"+urlLocation+"/admin/vacuum_volume_check", values) if err != nil { - fmt.Println("parameters:", values) + log.Println("parameters:", values) return err, false } var ret VacuumVolumeResult diff --git a/go/topology/topology_event_handling.go b/go/topology/topology_event_handling.go index fd2fe3bef..fff9b47c8 100644 --- a/go/topology/topology_event_handling.go +++ b/go/topology/topology_event_handling.go @@ -2,7 +2,7 @@ package topology import ( "code.google.com/p/weed-fs/go/storage" - "fmt" + "log" "math/rand" "time" ) @@ -28,10 +28,10 @@ func (t *Topology) StartRefreshWritableVolumes(garbageThreshold string) { t.SetVolumeCapacityFull(v) case dn := <-t.chanRecoveredDataNodes: t.RegisterRecoveredDataNode(dn) - fmt.Println("DataNode", dn, "is back alive!") + log.Println("DataNode", dn, "is back alive!") case dn := <-t.chanDeadDataNodes: t.UnRegisterDataNode(dn) - fmt.Println("DataNode", dn, "is dead!") + log.Println("DataNode", dn, "is dead!") } } }() @@ -48,7 +48,7 @@ func (t *Topology) SetVolumeCapacityFull(volumeInfo storage.VolumeInfo) bool { } func (t *Topology) UnRegisterDataNode(dn *DataNode) { for _, v := range dn.volumes { - fmt.Println("Removing Volume", v.Id, "from the dead volume server", dn) + log.Println("Removing Volume", v.Id, "from the dead volume server", dn) vl := t.GetVolumeLayout(v.RepType) vl.SetVolumeUnavailable(dn, v.Id) } diff --git a/go/topology/volume_layout.go b/go/topology/volume_layout.go index d8ed49b0b..7d7e3f662 100644 --- a/go/topology/volume_layout.go +++ b/go/topology/volume_layout.go @@ -3,7 +3,7 @@ package topology import ( "code.google.com/p/weed-fs/go/storage" "errors" - "fmt" + "log" "math/rand" ) @@ -54,7 +54,7 @@ func (vl *VolumeLayout) Lookup(vid storage.VolumeId) []*DataNode { func (vl *VolumeLayout) PickForWrite(count int, dataCenter string) (*storage.VolumeId, int, *VolumeLocationList, error) { len_writers := len(vl.writables) if len_writers <= 0 { - fmt.Println("No more writable volumes!") + log.Println("No more writable volumes!") return nil, 0, nil, errors.New("No more writable volumes!") } if dataCenter == "" { @@ -102,7 +102,7 @@ func (vl *VolumeLayout) GetActiveVolumeCount(dataCenter string) int { func (vl *VolumeLayout) removeFromWritable(vid storage.VolumeId) bool { for i, v := range vl.writables { if v == vid { - fmt.Println("Volume", vid, "becomes unwritable") + log.Println("Volume", vid, "becomes unwritable") vl.writables = append(vl.writables[:i], vl.writables[i+1:]...) return true } @@ -115,7 +115,7 @@ func (vl *VolumeLayout) setVolumeWritable(vid storage.VolumeId) bool { return false } } - fmt.Println("Volume", vid, "becomes writable") + log.Println("Volume", vid, "becomes writable") vl.writables = append(vl.writables, vid) return true } @@ -123,7 +123,7 @@ func (vl *VolumeLayout) setVolumeWritable(vid storage.VolumeId) bool { func (vl *VolumeLayout) SetVolumeUnavailable(dn *DataNode, vid storage.VolumeId) bool { if vl.vid2location[vid].Remove(dn) { if vl.vid2location[vid].Length() < vl.repType.GetCopyCount() { - fmt.Println("Volume", vid, "has", vl.vid2location[vid].Length(), "replica, less than required", vl.repType.GetCopyCount()) + log.Println("Volume", vid, "has", vl.vid2location[vid].Length(), "replica, less than required", vl.repType.GetCopyCount()) return vl.removeFromWritable(vid) } } |
