func equalType(x, y types.Type) error { if reflect.TypeOf(x) != reflect.TypeOf(y) { return fmt.Errorf("unequal kinds: %T vs %T", x, y) } switch x := x.(type) { case *types.Interface: y := y.(*types.Interface) // TODO(gri): enable separate emission of Embedded interfaces // and ExplicitMethods then use this logic. // if x.NumEmbeddeds() != y.NumEmbeddeds() { // return fmt.Errorf("unequal number of embedded interfaces: %d vs %d", // x.NumEmbeddeds(), y.NumEmbeddeds()) // } // for i := 0; i < x.NumEmbeddeds(); i++ { // xi := x.Embedded(i) // yi := y.Embedded(i) // if xi.String() != yi.String() { // return fmt.Errorf("mismatched %th embedded interface: %s vs %s", // i, xi, yi) // } // } // if x.NumExplicitMethods() != y.NumExplicitMethods() { // return fmt.Errorf("unequal methods: %d vs %d", // x.NumExplicitMethods(), y.NumExplicitMethods()) // } // for i := 0; i < x.NumExplicitMethods(); i++ { // xm := x.ExplicitMethod(i) // ym := y.ExplicitMethod(i) // if xm.Name() != ym.Name() { // return fmt.Errorf("mismatched %th method: %s vs %s", i, xm, ym) // } // if err := equalType(xm.Type(), ym.Type()); err != nil { // return fmt.Errorf("mismatched %s method: %s", xm.Name(), err) // } // } if x.NumMethods() != y.NumMethods() { return fmt.Errorf("unequal methods: %d vs %d", x.NumMethods(), y.NumMethods()) } for i := 0; i < x.NumMethods(); i++ { xm := x.Method(i) ym := y.Method(i) if xm.Name() != ym.Name() { return fmt.Errorf("mismatched %dth method: %s vs %s", i, xm, ym) } if err := equalType(xm.Type(), ym.Type()); err != nil { return fmt.Errorf("mismatched %s method: %s", xm.Name(), err) } } case *types.Array: y := y.(*types.Array) if x.Len() != y.Len() { return fmt.Errorf("unequal array lengths: %d vs %d", x.Len(), y.Len()) } if err := equalType(x.Elem(), y.Elem()); err != nil { return fmt.Errorf("array elements: %s", err) } case *types.Basic: y := y.(*types.Basic) if x.Kind() != y.Kind() { return fmt.Errorf("unequal basic types: %s vs %s", x, y) } case *types.Chan: y := y.(*types.Chan) if x.Dir() != y.Dir() { return fmt.Errorf("unequal channel directions: %d vs %d", x.Dir(), y.Dir()) } if err := equalType(x.Elem(), y.Elem()); err != nil { return fmt.Errorf("channel elements: %s", err) } case *types.Map: y := y.(*types.Map) if err := equalType(x.Key(), y.Key()); err != nil { return fmt.Errorf("map keys: %s", err) } if err := equalType(x.Elem(), y.Elem()); err != nil { return fmt.Errorf("map values: %s", err) } case *types.Named: y := y.(*types.Named) if x.String() != y.String() { return fmt.Errorf("unequal named types: %s vs %s", x, y) } case *types.Pointer: y := y.(*types.Pointer) if err := equalType(x.Elem(), y.Elem()); err != nil { return fmt.Errorf("pointer elements: %s", err) } case *types.Signature: y := y.(*types.Signature) if err := equalType(x.Params(), y.Params()); err != nil { return fmt.Errorf("parameters: %s", err) } if err := equalType(x.Results(), y.Results()); err != nil { return fmt.Errorf("results: %s", err) } if x.Variadic() != y.Variadic() { return fmt.Errorf("unequal varidicity: %t vs %t", x.Variadic(), y.Variadic()) } if (x.Recv() != nil) != (y.Recv() != nil) { return fmt.Errorf("unequal receivers: %s vs %s", x.Recv(), y.Recv()) } if x.Recv() != nil { // TODO(adonovan): fix: this assertion fires for interface methods. // The type of the receiver of an interface method is a named type // if the Package was loaded from export data, or an unnamed (interface) // type if the Package was produced by type-checking ASTs. // if err := equalType(x.Recv().Type(), y.Recv().Type()); err != nil { // return fmt.Errorf("receiver: %s", err) // } } case *types.Slice: y := y.(*types.Slice) if err := equalType(x.Elem(), y.Elem()); err != nil { return fmt.Errorf("slice elements: %s", err) } case *types.Struct: y := y.(*types.Struct) if x.NumFields() != y.NumFields() { return fmt.Errorf("unequal struct fields: %d vs %d", x.NumFields(), y.NumFields()) } for i := 0; i < x.NumFields(); i++ { xf := x.Field(i) yf := y.Field(i) if xf.Name() != yf.Name() { return fmt.Errorf("mismatched fields: %s vs %s", xf, yf) } if err := equalType(xf.Type(), yf.Type()); err != nil { return fmt.Errorf("struct field %s: %s", xf.Name(), err) } if x.Tag(i) != y.Tag(i) { return fmt.Errorf("struct field %s has unequal tags: %q vs %q", xf.Name(), x.Tag(i), y.Tag(i)) } } case *types.Tuple: y := y.(*types.Tuple) if x.Len() != y.Len() { return fmt.Errorf("unequal tuple lengths: %d vs %d", x.Len(), y.Len()) } for i := 0; i < x.Len(); i++ { if err := equalType(x.At(i).Type(), y.At(i).Type()); err != nil { return fmt.Errorf("tuple element %d: %s", i, err) } } } return nil }
func (p *exporter) typ(t types.Type) { if t == nil { log.Fatalf("gcimporter: nil type") } // Possible optimization: Anonymous pointer types *T where // T is a named type are common. We could canonicalize all // such types *T to a single type PT = *T. This would lead // to at most one *T entry in typIndex, and all future *T's // would be encoded as the respective index directly. Would // save 1 byte (pointerTag) per *T and reduce the typIndex // size (at the cost of a canonicalization map). We can do // this later, without encoding format change. // if we saw the type before, write its index (>= 0) if i, ok := p.typIndex[t]; ok { p.index('T', i) return } // otherwise, remember the type, write the type tag (< 0) and type data if trackAllTypes { if trace { p.tracef("T%d = {>\n", len(p.typIndex)) defer p.tracef("<\n} ") } p.typIndex[t] = len(p.typIndex) } switch t := t.(type) { case *types.Named: if !trackAllTypes { // if we don't track all types, track named types now p.typIndex[t] = len(p.typIndex) } p.tag(namedTag) p.pos(t.Obj()) p.qualifiedName(t.Obj()) p.typ(t.Underlying()) if !types.IsInterface(t) { p.assocMethods(t) } case *types.Array: p.tag(arrayTag) p.int64(t.Len()) p.typ(t.Elem()) case *types.Slice: p.tag(sliceTag) p.typ(t.Elem()) case *dddSlice: p.tag(dddTag) p.typ(t.elem) case *types.Struct: p.tag(structTag) p.fieldList(t) case *types.Pointer: p.tag(pointerTag) p.typ(t.Elem()) case *types.Signature: p.tag(signatureTag) p.paramList(t.Params(), t.Variadic()) p.paramList(t.Results(), false) case *types.Interface: p.tag(interfaceTag) p.iface(t) case *types.Map: p.tag(mapTag) p.typ(t.Key()) p.typ(t.Elem()) case *types.Chan: p.tag(chanTag) p.int(int(3 - t.Dir())) // hack p.typ(t.Elem()) default: log.Fatalf("gcimporter: unexpected type %T: %s", t, t) } }
// hashFor computes the hash of t. func (h Hasher) hashFor(t types.Type) uint32 { // See Identical for rationale. switch t := t.(type) { case *types.Basic: return uint32(t.Kind()) case *types.Array: return 9043 + 2*uint32(t.Len()) + 3*h.Hash(t.Elem()) case *types.Slice: return 9049 + 2*h.Hash(t.Elem()) case *types.Struct: var hash uint32 = 9059 for i, n := 0, t.NumFields(); i < n; i++ { f := t.Field(i) if f.Anonymous() { hash += 8861 } hash += hashString(t.Tag(i)) hash += hashString(f.Name()) // (ignore f.Pkg) hash += h.Hash(f.Type()) } return hash case *types.Pointer: return 9067 + 2*h.Hash(t.Elem()) case *types.Signature: var hash uint32 = 9091 if t.Variadic() { hash *= 8863 } return hash + 3*h.hashTuple(t.Params()) + 5*h.hashTuple(t.Results()) case *types.Interface: var hash uint32 = 9103 for i, n := 0, t.NumMethods(); i < n; i++ { // See go/types.identicalMethods for rationale. // Method order is not significant. // Ignore m.Pkg(). m := t.Method(i) hash += 3*hashString(m.Name()) + 5*h.Hash(m.Type()) } return hash case *types.Map: return 9109 + 2*h.Hash(t.Key()) + 3*h.Hash(t.Elem()) case *types.Chan: return 9127 + 2*uint32(t.Dir()) + 3*h.Hash(t.Elem()) case *types.Named: // Not safe with a copying GC; objects may move. return uint32(reflect.ValueOf(t.Obj()).Pointer()) case *types.Tuple: return h.hashTuple(t) } panic(t) }