Commit 95d0183c authored by Kirill Smelkov's avatar Kirill Smelkov

.

parent d77ccb51
......@@ -136,7 +136,7 @@ func (nt *NodeTable) GetByLink(link *NodeLink) *Node {
}
// XXX doc
func (nt *NodeTable) SetNodeState(node *Node, state neo.NodeState) {
func (nt *NodeTable) SetNodeState(node *Node, state NodeState) {
node.NodeState = state
nt.notify(node.NodeInfo)
}
......
......@@ -22,7 +22,7 @@ package server
import (
"context"
"errors"
stderrors "errors"
"fmt"
// "math"
"sync"
......@@ -31,6 +31,8 @@ import (
"lab.nexedi.com/kirr/neo/go/neo"
"lab.nexedi.com/kirr/neo/go/zodb"
"lab.nexedi.com/kirr/neo/go/xcommon/log"
"lab.nexedi.com/kirr/neo/go/xcommon/xcontext"
"lab.nexedi.com/kirr/neo/go/xcommon/xnet"
"lab.nexedi.com/kirr/go123/xerr"
......@@ -150,7 +152,7 @@ func (m *Master) Run(ctx context.Context) error {
if err != nil {
return err // XXX err ctx
}
m.logf("serving on %s ...", l.Addr())
log.Infof(ctx, "serving on %s ...", l.Addr())
m.node.MasterAddr = l.Addr().String()
......@@ -194,7 +196,7 @@ func (m *Master) Run(ctx context.Context) error {
// runMain is the process which implements main master cluster management logic: node tracking, cluster
// state updates, scheduling data movement between storage nodes etc
func (m *Master) runMain(ctx context.Context) (err error) {
//defer xerr.Context(&err, "master: run")
defer running(&ctx, "run")(&err) // XXX needed?
// NOTE Run's goroutine is the only mutator of nodeTab, partTab and other cluster state
......@@ -207,7 +209,7 @@ func (m *Master) runMain(ctx context.Context) (err error) {
// a command came to us to start the cluster.
err := m.recovery(ctx)
if err != nil {
m.log(err)
log.Error(ctx, err)
return err // recovery cancelled
}
......@@ -215,14 +217,14 @@ func (m *Master) runMain(ctx context.Context) (err error) {
// case previously it was unclean shutdown.
err = m.verify(ctx)
if err != nil {
m.log(err)
log.Error(ctx, err)
continue // -> recovery
}
// provide service as long as partition table stays operational
err = m.service(ctx)
if err != nil {
m.log(err)
log.Error(ctx, err)
continue // -> recovery
}
......@@ -231,7 +233,7 @@ func (m *Master) runMain(ctx context.Context) (err error) {
// XXX shutdown ?
}
return errors.WithMessage(ctx.Err(), "master: run")
return ctx.Err()
}
......@@ -246,7 +248,7 @@ func (m *Master) runMain(ctx context.Context) (err error) {
// storRecovery is result of 1 storage node passing recovery phase
type storRecovery struct {
node *Node
node *neo.Node
partTab neo.PartitionTable
// XXX + backup_tid, truncate_tid ?
err error
......@@ -260,12 +262,8 @@ type storRecovery struct {
func (m *Master) recovery(ctx context.Context) (err error) {
defer running(&ctx, "recovery")(&err)
ctx = task.Running(ctx, "recovery")
defer task.ErrContext(&err, ctx)
log.Infof(ctx, "") // XXX automatically log in task.Running?
m.setClusterState(neo.ClusterRecovering)
rctx, rcancel := context.WithCancel(ctx)
ctx, rcancel := context.WithCancel(ctx)
defer rcancel()
recovery := make(chan storRecovery)
......@@ -278,7 +276,7 @@ func (m *Master) recovery(ctx context.Context) (err error) {
wg.Add(1)
go func() {
defer wg.Done()
storCtlRecovery(rctx, stor, recovery)
storCtlRecovery(ctx, stor, recovery)
}()
}
}
......@@ -296,7 +294,7 @@ loop:
defer wg.Done()
if node == nil {
m.reject(n.conn, resp)
m.reject(ctx, n.conn, resp)
return
}
......@@ -308,7 +306,7 @@ loop:
}
// start recovery
storCtlRecovery(rctx, node, recovery)
storCtlRecovery(ctx, node, recovery)
}()
/*
......@@ -325,18 +323,18 @@ loop:
// ptid ↑ and if so we should take partition table from there
case r := <-recovery:
if r.err != nil {
logf(ctx, "%v", r.err)
log.Error(ctx, r.err)
if !xcontext.Canceled(errors.Cause(r.err)) {
logf(ctx, "%v: closing link", r.node.Link)
log.Infof(ctx, "%v: closing link", r.node.Link)
// close stor link / update .nodeTab
err := r.node.Link.Close()
if err != nil {
logf(ctx, "master: %v", err)
log.Error(ctx, err)
}
m.nodeTab.SetNodeState(r.node, DOWN)
m.nodeTab.SetNodeState(r.node, neo.DOWN)
}
} else {
......@@ -393,7 +391,7 @@ loop:
for {
select {
// XXX <-m.nodeLeave ?
case <-recovery:
case r := <-recovery:
// we do not care errors here - they are either cancelled or IO errors
// we just log them and return - in case it is IO error
// on link it will be caught on next send/recv XXX
......@@ -421,9 +419,9 @@ func storCtlRecovery(ctx context.Context, stor *neo.Node, res chan storRecovery)
// on error provide feedback to storRecovery chan
res <- storRecovery{node: stor, err: err}
}()
defer xerr.Contextf(&err, "%s: stor recovery", stor.link)
defer runningf(&ctx, "%s: stor recovery", stor.Link)(&err)
conn, err := stor.link.NewConn()
conn, err := stor.Link.NewConn()
if err != nil {
return
}
......@@ -468,8 +466,8 @@ func storCtlRecovery(ctx context.Context, stor *neo.Node, res chan storRecovery)
}
var errStopRequested = errors.New("stop requested")
var errClusterDegraded = errors.New("cluster became non-operatonal")
var errStopRequested = stderrors.New("stop requested")
var errClusterDegraded = stderrors.New("cluster became non-operatonal")
// Cluster Verification (data recovery)
......@@ -492,8 +490,7 @@ var errClusterDegraded = errors.New("cluster became non-operatonal")
//
// prerequisite for start: .partTab is operational wrt .nodeTab
func (m *Master) verify(ctx context.Context) (err error) {
m.log("verify")
defer xerr.Context(&err, "master: verify")
defer running(&ctx, "verify")(&err)
m.setClusterState(neo.ClusterVerifying)
vctx, vcancel := context.WithCancel(ctx)
......@@ -512,7 +509,7 @@ func (m *Master) verify(ctx context.Context) (err error) {
for _, stor := range m.nodeTab.StorageList() {
if stor.NodeState > neo.DOWN { // XXX state cmp ok ? XXX or stor.Link != nil ?
inprogress++
go storCtlVerify(vctx, stor.Link, verify)
go storCtlVerify(vctx, stor, verify)
}
}
......@@ -520,7 +517,9 @@ loop:
for inprogress > 0 {
select {
case n := <-m.nodeCome:
node, ok := m.identify(n, /* XXX only accept storages -> known ? RUNNING : PENDING */)
node, resp := m.identify(n, /* XXX only accept storages -> known ? RUNNING : PENDING */)
// XXX handle resp ^^^ like in recover
_, ok := resp.(*neo.AcceptIdentification)
if !ok {
break
}
......@@ -528,10 +527,10 @@ loop:
// new storage arrived - start verification on it too
// XXX ok? or it must first go through recovery check?
inprogress++
go storCtlVerify(vctx, node.Link, verify)
go storCtlVerify(vctx, node, verify)
case n := <-m.nodeLeave:
m.nodeTab.UpdateLinkDown(n.link)
m.nodeTab.SetNodeState(n.node, neo.DOWN)
// if cluster became non-operational - we cancel verification
if !m.partTab.OperationalWith(&m.nodeTab) {
......@@ -548,11 +547,10 @@ loop:
inprogress--
if v.err != nil {
m.logf("verify: %v", v.err)
log.Error(ctx, v.err)
// mark storage as non-working in nodeTab
// FIXME better -> v.node.setState(DOWN) ?
m.nodeTab.UpdateLinkDown(v.link)
m.nodeTab.SetNodeState(v.node, neo.DOWN)
// check partTab is still operational
// if not -> cancel to go back to recovery
......@@ -597,7 +595,7 @@ loop:
// storVerify is result of a storage node passing verification phase
type storVerify struct {
node *Node
node *neo.Node
lastOid zodb.Oid
lastTid zodb.Tid
// link *neo.NodeLink // XXX -> Node
......@@ -605,20 +603,20 @@ type storVerify struct {
}
// storCtlVerify drives a storage node during cluster verifying (= starting) state
func storCtlVerify(ctx context.Context, link *neo.NodeLink, res chan storVerify) {
func storCtlVerify(ctx context.Context, stor *neo.Node, res chan storVerify) {
// XXX link.Close on err
// XXX cancel on ctx
var err error
defer func() {
if err != nil {
res <- storVerify{link: link, err: err}
res <- storVerify{node: stor, err: err}
}
}()
defer xerr.Contextf(&err, "%s: verify", link)
defer runningf(&ctx, "%s: stor verify", stor.Link)(&err)
// FIXME stub
conn, _ := link.NewConn()
conn, _ := stor.Link.NewConn()
// XXX NotifyPT (so storages save locally recovered PT)
......@@ -641,7 +639,7 @@ func storCtlVerify(ctx context.Context, link *neo.NodeLink, res chan storVerify)
}
// send results to driver
res <- storVerify{link: link, lastOid: last.LastOid, lastTid: last.LastTid}
res <- storVerify{node: stor, lastOid: last.LastOid, lastTid: last.LastTid}
}
......@@ -661,8 +659,7 @@ func storCtlVerify(ctx context.Context, link *neo.NodeLink, res chan storVerify)
//
// prerequisite for start: .partTab is operational wrt .nodeTab and verification passed
func (m *Master) service(ctx context.Context) (err error) {
m.log("service")
defer xerr.Context(&err, "master: service")
defer running(&ctx, "service")(&err)
// XXX we also need to tell storages StartOperation first
m.setClusterState(neo.ClusterRunning)
......@@ -674,10 +671,11 @@ loop:
select {
// a node connected and requests identification
case n := <-m.nodeCome:
node, resp, ok := m.identify(n, /* XXX accept everyone */)
state := m.ClusterState
node, resp := m.identify(n, /* XXX accept everyone */)
//state := m.clusterState
_ = node
_ = resp
/*
wg.Add(1)
go func() {
......@@ -784,7 +782,7 @@ loop:
*/
case n := <-m.nodeLeave:
m.nodeTab.UpdateLinkDown(n.link)
m.nodeTab.SetNodeState(n.node, neo.DOWN)
// if cluster became non-operational - cancel service
if !m.partTab.OperationalWith(&m.nodeTab) {
......@@ -795,12 +793,13 @@ loop:
// XXX what else ? (-> txn control at least)
/*
case ech := <-m.ctlStart:
switch m.ClusterState {
case ClusterVerifying, ClusterRunning:
switch m.clusterState {
case neo.ClusterVerifying, neo.ClusterRunning:
ech <- nil // we are already started
case ClusterRecovering:
case neo.ClusterRecovering:
if m.partTab.OperationalWith(&m.nodeTab) {
// reply "ok to start" after whole recovery finishes
......@@ -820,6 +819,7 @@ loop:
// XXX case ClusterStopping:
}
*/
case ech := <-m.ctlStop:
close(ech) // ok
......@@ -907,13 +907,16 @@ func (m *Master) identify(n nodeCome) (node *neo.Node, resp neo.Msg) {
}
// reject sends rejective identification response and closes associated link
func (m *Master) reject(conn *neo.Conn, resp neo.Msg) {
func (m *Master) reject(ctx context.Context, conn *neo.Conn, resp neo.Msg) {
// XXX cancel on ctx?
// XXX log?
err1 := conn.Send(resp)
err2 := conn.Close()
err3 := conn.Link().Close()
log xerr.Merge(err1, err2, err3)
err := xerr.Merge(err1, err2, err3)
if err != nil {
log.Error(ctx, "reject:", err)
}
}
// accept sends acceptive identification response and closes conn
......@@ -1255,35 +1258,3 @@ func (m *Master) ServeAdmin(ctx context.Context, conn *neo.Conn) {
func (m *Master) ServeMaster(ctx context.Context, conn *neo.Conn) {
// TODO (for elections)
}
// ---- utils ----
// -> Logger
//func (m *Master) log(v interface{}) {
// XXX
//}
func (m *Master) logf(format string, argv ...interface{}) {
// TODO support custom log.Logger
log.Output(2, fmt.Sprintf("master(%v): " + format, append([]string{m.MasterAddr}, argv...)...))
}
func (m *Master) vlogf(format string, argv ...interface{}) {
// XXX -> verbose settings
if false {
return
}
m.log(format, argv)
}
func (m *Master) errctx(errp *error, context string) {
if *errp == nil {
return
}
xerr.Contextf(errp, "master(%v): %s", m.node.MasterAddr, context)
}
......@@ -79,7 +79,8 @@ func Serve(ctx context.Context, l *neo.Listener, srv Server) error {
}
*/
/*
// FIXME kill vvv
///*
// ----------------------------------------
// XXX goes away? (we need a func to make sure to recv RequestIdentification
......@@ -129,4 +130,4 @@ func IdentifyPeer(link *neo.NodeLink, myNodeType neo.NodeType) (nodeInfo neo.Req
return req, nil
}
*/
//*/
......@@ -28,6 +28,7 @@ import (
"lab.nexedi.com/kirr/neo/go/neo"
"lab.nexedi.com/kirr/neo/go/zodb"
"lab.nexedi.com/kirr/neo/go/xcommon/log"
"lab.nexedi.com/kirr/neo/go/xcommon/xnet"
"lab.nexedi.com/kirr/neo/go/xcommon/xcontext"
......@@ -92,7 +93,7 @@ func (stor *Storage) Run(ctx context.Context) error {
if err != nil {
return err // XXX err ctx
}
stor.logf("serving on %s ...", l.Addr())
log.Infof(ctx, "serving on %s ...", l.Addr())
// start serving incoming connections
wg := sync.WaitGroup{}
......@@ -140,13 +141,13 @@ func (stor *Storage) Run(ctx context.Context) error {
// it tries to persist master link reconnecting as needed
//
// it always returns an error - either due to cancel or command from master to shutdown
func (stor *Storage) talkMaster(ctx context.Context) error {
// XXX errctx
func (stor *Storage) talkMaster(ctx context.Context) (err error) {
defer runningf(&ctx, "talk master(%v)", stor.node.MasterAddr)(&err)
for {
stor.vlogf("master(%v): connecting ...", stor.node.MasterAddr)
log.Info(ctx, "connecting ...")
err := stor.talkMaster1(ctx)
stor.logf("master(%v): %v", stor.node.MasterAddr, err)
log.Error(ctx, err)
// TODO if err = shutdown -> return
......@@ -187,7 +188,7 @@ func (stor *Storage) talkMaster1(ctx context.Context) (err error) {
// XXX -> node.Dial ?
if accept.YourNodeUUID != stor.node.MyInfo.NodeUUID {
stor.logf("%v: master told us to have UUID=%v", Mlink, accept.YourNodeUUID)
log.Infof(ctx, "%v: master told us to have UUID=%v", Mlink, accept.YourNodeUUID)
stor.node.MyInfo.NodeUUID = accept.YourNodeUUID
}
......@@ -225,13 +226,13 @@ func (stor *Storage) talkMaster1(ctx context.Context) (err error) {
// let master initialize us. If successful this ends with StartOperation command.
err = stor.m1initialize(ctx, Mconn)
if err != nil {
stor.logf("%v: master: %v", Mconn, err)
log.Error(ctx, err)
continue // retry initializing
}
// we got StartOperation command. Let master drive us during servicing phase.
err = stor.m1serve(ctx, Mconn)
stor.logf("%v: master: %v", Mconn, err)
log.Error(ctx, err)
continue // retry from initializing
}
......@@ -250,7 +251,7 @@ func (stor *Storage) talkMaster1(ctx context.Context) (err error) {
// - nil: initialization was ok and a command came from master to start operation
// - !nil: initialization was cancelled or failed somehow
func (stor *Storage) m1initialize(ctx context.Context, Mconn *neo.Conn) (err error) {
defer xerr.Context(&err, "init")
defer runningf(&ctx, "init %v", Mconn)(&err)
for {
msg, err := Mconn.Recv()
......@@ -319,7 +320,7 @@ func (stor *Storage) m1initialize(ctx context.Context, Mconn *neo.Conn) (err err
// either due to master commanding us to stop, or context cancel or some other
// error.
func (stor *Storage) m1serve(ctx context.Context, Mconn *neo.Conn) (err error) {
defer xerr.Context(&err, "serve")
defer runningf(&ctx, "serve %v", Mconn)(&err)
// refresh stor.opCtx and cancel it when we finish so that client
// handlers know they need to stop operating as master told us to do so.
......@@ -359,7 +360,7 @@ func (stor *Storage) m1serve(ctx context.Context, Mconn *neo.Conn) (err error) {
// ServeLink serves incoming node-node link connection
// XXX +error return?
func (stor *Storage) ServeLink(ctx context.Context, link *neo.NodeLink) {
stor.logf("%s: serving new node", link)
log.Infof(ctx, "%s: serving new node", link) // XXX -> running?
// close link when either cancelling or returning (e.g. due to an error)
// ( when cancelling - link.Close will signal to all current IO to
......@@ -374,7 +375,7 @@ func (stor *Storage) ServeLink(ctx context.Context, link *neo.NodeLink) {
// XXX ret err = ctx.Err()
case <-retch:
}
stor.logf("%v: closing link", link)
log.Info(ctx, "%v: closing link", link)
link.Close() // XXX err
}()
......@@ -382,7 +383,7 @@ func (stor *Storage) ServeLink(ctx context.Context, link *neo.NodeLink) {
// XXX only accept when operational (?)
nodeInfo, err := IdentifyPeer(link, neo.STORAGE)
if err != nil {
stor.logf("%v", err)
log.Error(ctx, err)
return
}
......@@ -393,7 +394,7 @@ func (stor *Storage) ServeLink(ctx context.Context, link *neo.NodeLink) {
default:
// XXX vvv should be reply to peer
stor.logf("%v: unexpected peer type: %v", link, nodeInfo.NodeType)
log.Errorf(ctx, "%v: unexpected peer type: %v", link, nodeInfo.NodeType)
return
}
......@@ -401,7 +402,7 @@ func (stor *Storage) ServeLink(ctx context.Context, link *neo.NodeLink) {
for {
conn, err := link.Accept()
if err != nil {
stor.logf("%v", err) // XXX err ctx
log.Error(ctx, err)
break
}
......@@ -427,7 +428,7 @@ func (stor *Storage) withWhileOperational(ctx context.Context) (context.Context,
// the connection is closed when serveClient returns
// XXX +error return?
func (stor *Storage) serveClient(ctx context.Context, conn *neo.Conn) {
stor.logf("%s: serving new client conn", conn)
log.Infof(ctx, "%s: serving new client conn", conn) // XXX -> running?
// rederive ctx to be also cancelled if M tells us StopOperation
ctx, cancel := stor.withWhileOperational(ctx)
......@@ -458,9 +459,9 @@ func (stor *Storage) serveClient(ctx context.Context, conn *neo.Conn) {
case err = <-done:
}
stor.logf("%v: %v", conn, err)
log.Infof(ctx, "%v: %v", conn, err)
// XXX vvv -> defer ?
stor.logf("%v: closing client conn", conn)
log.Infof(ctx, "%v: closing client conn", conn)
conn.Close() // XXX err
}
......
......@@ -19,3 +19,39 @@
package server
// misc utilities
import (
"context"
"fmt"
"lab.nexedi.com/kirr/neo/go/xcommon/task"
"lab.nexedi.com/kirr/neo/go/xcommon/log"
)
// running is syntatic sugar to push new task to operational stack, log it and
// adjust error return with task prefix.
//
// use like this:
//
// defer running(&ctx, "my task")(&err)
func running(ctxp *context.Context, name string) func(*error) {
return _running(ctxp, name)
}
// runningf is running cousing with formatting support
func runningf(ctxp *context.Context, format string, argv ...interface{}) func(*error) {
return _running(ctxp, fmt.Sprintf(format, argv...))
}
func _running(ctxp *context.Context, name string) func(*error) {
ctx := task.Running(*ctxp, name)
*ctxp = ctx
log.Depth(2).Info(ctx) // XXX level = ok?
return func(errp *error) {
// NOTE not *ctxp here - as context pointed by ctxp could be
// changed when this deferred function is run
task.ErrContext(errp, ctx)
// XXX also log task stop here?
}
}
......@@ -23,30 +23,79 @@
package log
import (
"context"
"fmt"
"github.com/golang/glog"
"lab.nexedi.com/kirr/neo/go/xcommon/task"
)
/*
// taskPrefix returns prefix associated to current operational task stack to put to logs
func taskPrefix(ctx) string {
func taskPrefix(ctx context.Context) string {
s := task.Current(ctx).String()
if s != "" {
s += ": "
}
return s
}
*/
// withTask prepends string describing current operational task stack to argv and returns it
// handy to use this way:
//
// func info(ctx, argv ...interface{}) {
// glog.Info(withTask(ctx, argv...)...)
// }
//
// see https://golang.org/issues/21388
func withTask(ctx context.Context, argv ...interface{}) []interface{} {
task := task.Current(ctx).String()
if task == "" {
return argv
}
if len(argv) != 0 {
task += ":"
}
return append([]interface{}{task}, argv...)
}
type Depth int
func (d Depth) Info(ctx context.Context, argv ...interface{}) {
// XXX avoid task formatting if logging severity disabled
glog.InfoDepth(int(d+1), withTask(ctx, argv...)...)
}
func (d Depth) Infof(ctx context.Context, format string, argv ...interface{}) {
// XXX avoid formatting if logging severity disables info
glog.InfoDepth(d+1, taskPrefix(ctx) + fmt.Sprintf(format, argv)
// XXX avoid formatting if logging severity disabled
glog.InfoDepth(int(d+1), withTask(ctx, fmt.Sprintf(format, argv))...)
}
func (d Depth) Error(ctx context.Context, argv ...interface{}) {
glog.ErrorDepth(int(d+1), withTask(ctx, argv...)...)
}
func (d Depth) Errorf(ctx context.Context, format string, argv ...interface{}) {
glog.ErrorDepth(int(d+1), withTask(ctx, fmt.Sprintf(format, argv))...)
}
func Info(ctx context.Context, argv ...interface{}) { Depth(1).Info(ctx, argv...) }
func Error(ctx context.Context, argv ...interface{}) { Depth(1).Error(ctx, argv...) }
func Infof(ctx context.Context, format string, argv ...interface{}) {
Depth(1).Infof(ctx, format, argv)
Depth(1).Infof(ctx, format, argv...)
}
func Errorf(ctx context.Context, format string, argv ...interface{}) {
Depth(1).Errorf(ctx, format, argv...)
}
......
......@@ -22,6 +22,9 @@ package task
import (
"context"
"fmt"
"lab.nexedi.com/kirr/go123/xerr"
)
// Task represents currently running operation
......@@ -34,13 +37,18 @@ type taskKey struct{}
// Running creates new task and returns new context with that task set to current
func Running(ctx context.Context, name string) context.Context {
context.WithValue(&Task{Parent: Current(ctx), Name: name})
return context.WithValue(ctx, taskKey{}, &Task{Parent: Current(ctx), Name: name})
}
// Runningf is Running cousing with formatting support
func Runningf(ctx context.Context, format string, argv ...interface{}) context.Context {
return Running(ctx, fmt.Sprintf(format, argv...))
}
// Current returns current task represented by context.
// if there is no current task - it returns nil.
func Current(ctx Context) *Task {
task, _ := ctx.Value(taskKey).(*Task)
func Current(ctx context.Context) *Task {
task, _ := ctx.Value(taskKey{}).(*Task)
return task
}
......@@ -49,16 +57,16 @@ func Current(ctx Context) *Task {
//
// func myfunc(ctx, ...) (..., err error) {
// ctx = task.Running("doing something")
// defer task.ErrContext(&err, ctx)
// ...
// defer task.ErrContext(&err, ctx)
// ...
//
// Please see lab.nexedi.com/kirr/go123/xerr.Context for semantic details
func ErrContext(errp *error, ctx Context) {
func ErrContext(errp *error, ctx context.Context) {
task := Current(ctx)
if task == nil {
return
}
return xerr.Context(errp, task.Name)
xerr.Context(errp, task.Name)
}
// String returns string representing whole operational stack.
......@@ -68,11 +76,11 @@ func ErrContext(errp *error, ctx Context) {
//
// nil Task is represented as ""
func (t *Task) String() string {
if o == nil {
if t == nil {
return ""
}
prefix := Parent.String()
prefix := t.Parent.String()
if prefix != "" {
prefix += ": "
}
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment